Airflow is well-suited for data engineering pipelines, creating scheduled workflows, and working with various data sources. You can implement almost any kind of DAG for any use case using the different operators or enforce your operator using the Python operator with ease. The MLOps feature of Airflow can be enhanced to match MLFlow-like features, making Airflow the go-to solution for all workloads, from data science to data engineering.
In the context of Internet of Things (IoT) applications, IBM MQ plays a pivotal role in managing the substantial data streams emanating from interconnected devices. Its primary function is to guarantee the dependable transmission and processing of data, catering to a diverse range of IoT use cases, including but not limited to smart city initiatives, healthcare monitoring systems, and industrial automation solutions. In the telecommunications sector, IBM MQ is employed for message routing, call detail record (CDR) processing, and network management to ensure real-time data exchange and fault tolerance. When managing the supply chain and logistics, IBM MQ is used to ensure timely and accurate communication between different entities, including suppliers, warehouses, and transportation providers. IBM MQ can be cost-prohibitive for smaller organizations due to licensing and maintenance costs. In such cases, open-source or lightweight messaging solutions may be more appropriate. For scenarios requiring extremely low-latency, real-time data exchange, and high throughput, other messaging technologies, like Apache Kafka, may be more suitable due to their specialized design for such use cases.
Apache Airflow is one of the best Orchestration platforms and a go-to scheduler for teams building a data platform or pipelines.
Apache Airflow supports multiple operators, such as the Databricks, Spark, and Python operators. All of these provide us with functionality to implement any business logic.
Apache Airflow is highly scalable, and we can run a large number of DAGs with ease. It provided HA and replication for workers. Maintaining airflow deployments is very easy, even for smaller teams, and we also get lots of metrics for observability.
The documentation is very clear,It is understandable and the support helps to configure it in the best way.
Server guidelines make it possible to get the most out of work management. It's broad, we can work with different operating systems, I really recommend using linux.
It is highly compatible with systems, brockers, applications, and data accumulation programs, it is possible to configure everything so that after the installation of programs, they can communicate with each other and then throw data to an external program that accumulates it and represents in clear details of steps to follow and make business decisions.
UI/Dashboard can be updated to be customisable, and jobs summary in groups of errors/failures/success, instead of each job, so that a summary of errors can be used as a starting point for reviewing them.
Navigation - It's a bit dated. Could do with more modern web navigation UX. i.e. sidebars navigation instead of browser back/forward.
Again core functional reorg in terms of UX. Navigation can be improved for core functions as well, instead of discovery.
There is limitation on number of svrconn connections you can have to MQ on the mainframe which has been an major issue for us. This has been an issue for us for over 4 years and still no fix although I am aware IBM have been working on a solution over the last year.
When upgrading to MQ V9.3 on our MQ appliances there is no fall-back option. This was the same for MQ V9.2 upgrade from MQ V9.0. For production upgrades this I believe is not acceptable.
AMS is not supplied as part of the standard mainframe MQ licence. You need an extra licence. IBM tell customers how important security and protecting data is yet they still want to charge for this software. The cost of MQ on the mainframe is not cheap so I would expect AMS to be part of the base product.
For its capability to connect with multicloud environments. Access Control management is something that we don't get in all the schedulers and orchestrators. But although it provides so many flexibility and options to due to python , some level of knowledge of python is needed to be able to build workflows.
I give it a nine because it has significantly improved my team's data reliability and operational efficiency. Its great security features give us peace of mind, knowing our sensitive data is well protected. While the setup might initially be complex, I believe the long-term benefits far outweigh this hurdle.
The messages are delivered instantly with this software and it integrates with our technology stack, in terms of availability we only had one failure when we were doing some testing and integration with third parties, the features of this software make it always available and its deployment is easy for the company, it does not generate expenses due to failures
There are very specific things that must be elevated to more specialized areas of support, but the common support is very agile when receiving questions or when we leave concerns in real time. I recommend the support of the program in this regard.
Multiple DAGs can be orchestrated simultaneously at varying times, and runs can be reproduced or replicated with relative ease. Overall, utilizing Apache Airflow is easier to use than other solutions now on the market. It is simple to integrate in Apache Airflow, and the workflow can be monitored and scheduling can be done quickly using Apache Airflow. We advocate using this tool for automating the data pipeline or process.
We found IBM MQ very easy to get started and quick to learn by the new users with a short learning curve and seamlessly integrates with IBM products, and quick to perform self-service analytics and make informed business decisions. IBM MQ is also very straightforward in creating simple and best reports, which are very profitable and productive.
Impact Depends on number of workflows. If there are lot of workflows then it has a better usecase as the implementation is justified as it needs resources , dedicated VMs, Database that has a cost