Apache Camel vs. Apache Spark

Overview
ProductRatingMost Used ByProduct SummaryStarting Price
Apache Camel
Score 6.5 out of 10
N/A
Apache Camel is an open source integration platform.N/A
Apache Spark
Score 8.6 out of 10
N/A
N/AN/A
Pricing
Apache CamelApache Spark
Editions & Modules
No answers on this topic
No answers on this topic
Offerings
Pricing Offerings
Apache CamelApache Spark
Free Trial
NoNo
Free/Freemium Version
NoNo
Premium Consulting/Integration Services
NoNo
Entry-level Setup FeeNo setup feeNo setup fee
Additional Details
More Pricing Information
Community Pulse
Apache CamelApache Spark
Considered Both Products
Apache Camel

No answer on this topic

Apache Spark
Chose Apache Spark
1. Apache Spark is almost 100 % faster than Hadoop.
2. Apache Spark is more stable than Amazon EMR.
3. The end to end distributed machine library is more robust in Apache Spark.
Chose Apache Spark
I prefer Apache Spark compared to Hadoop, since in my experience Spark has more usability and comes equipped with simple APIs for Scala, Python, Java and Spark SQL, as well as provides feedback in REPL format on the commands. At the same time, Apache Spark seems to have the …
Top Pros
Top Cons
Best Alternatives
Apache CamelApache Spark
Small Businesses

No answers on this topic

No answers on this topic

Medium-sized Companies
Anypoint Platform
Anypoint Platform
Score 8.1 out of 10
Cloudera Manager
Cloudera Manager
Score 9.7 out of 10
Enterprises
TIBCO B2B Integration Solution
TIBCO B2B Integration Solution
Score 8.0 out of 10
IBM Analytics Engine
IBM Analytics Engine
Score 8.8 out of 10
All AlternativesView all alternativesView all alternatives
User Ratings
Apache CamelApache Spark
Likelihood to Recommend
7.8
(11 ratings)
9.9
(24 ratings)
Likelihood to Renew
-
(0 ratings)
10.0
(1 ratings)
Usability
-
(0 ratings)
10.0
(3 ratings)
Support Rating
-
(0 ratings)
8.7
(4 ratings)
User Testimonials
Apache CamelApache Spark
Likelihood to Recommend
Apache
Message brokering across different systems, with transactionality and the ability to have fine tuned control over what happens using Java (or other languages), instead of a heavy, proprietary languages. One situation that it doesn't fit very well (as far as I have experienced) is when your workflow requires significant data mapping. While possible when using Java tooling, some other visual data mapping tools in other integration frameworks are easier to work with.
Read full review
Apache
Well suited: To most of the local run of datasets and non-prod systems - scalability is not a problem at all. Including data from multiple types of data sources is an added advantage. MLlib is a decently nice built-in library that can be used for most of the ML tasks. Less appropriate: We had to work on a RecSys where the music dataset that we used was around 300+Gb in size. We faced memory-based issues. Few times we also got memory errors. Also the MLlib library does not have support for advanced analytics and deep-learning frameworks support. Understanding the internals of the working of Apache Spark for beginners is highly not possible.
Read full review
Pros
Apache
  • Camel has an easy learning curve. It is fairly well documented and there are about 5-6 books on Camel.
  • There is a large user group and blogs devoted to all things Camel and the developers of Camel provide quick answers and have also been very quick to patch Camel, when bugs are reported.
  • Camel integrates well with well known frameworks like Spring, and other middleware products like Apache Karaf and Servicemix.
  • There are over 150 components for the Camel framework that help integrate with diverse software platforms.
  • Camel is also good for creating microservices.
Read full review
Apache
  • Apache Spark makes processing very large data sets possible. It handles these data sets in a fairly quick manner.
  • Apache Spark does a fairly good job implementing machine learning models for larger data sets.
  • Apache Spark seems to be a rapidly advancing software, with the new features making the software ever more straight-forward to use.
Read full review
Cons
Apache
  • didn't work well when our developers tried to transform heavy data sets
  • Apache Camel's whole logic is based on java so team needs to have a great skill set in java
  • if there are a handful of workflows then Apache Camel's full potential can't be realized
Read full review
Apache
  • Memory management. Very weak on that.
  • PySpark not as robust as scala with spark.
  • spark master HA is needed. Not as HA as it should be.
  • Locality should not be a necessity, but does help improvement. But would prefer no locality
Read full review
Likelihood to Renew
Apache
No answers on this topic
Apache
Capacity of computing data in cluster and fast speed.
Read full review
Usability
Apache
No answers on this topic
Apache
The only thing I dislike about spark's usability is the learning curve, there are many actions and transformations, however, its wide-range of uses for ETL processing, facility to integrate and it's multi-language support make this library a powerhouse for your data science solutions. It has especially aided us with its lightning-fast processing times.
Read full review
Support Rating
Apache
No answers on this topic
Apache
1. It integrates very well with scala or python. 2. It's very easy to understand SQL interoperability. 3. Apache is way faster than the other competitive technologies. 4. The support from the Apache community is very huge for Spark. 5. Execution times are faster as compared to others. 6. There are a large number of forums available for Apache Spark. 7. The code availability for Apache Spark is simpler and easy to gain access to. 8. Many organizations use Apache Spark, so many solutions are available for existing applications.
Read full review
Alternatives Considered
Apache
If you are looking for a Java-based open source low cost equivalent to webMethods or Azure Logic Apps, Apache Camel is an excellent choice as it is mature and widely deployed, and included in many vendored Java application servers too such as Redhat JBoss EAP. Apache Camel is lacking on the GUI tooling side compared to commercial products such as webMethods or Azure Logic Apps.
Read full review
Apache
All the above systems work quite well on big data transformations whereas Spark really shines with its bigger API support and its ability to read from and write to multiple data sources. Using Spark one can easily switch between declarative versus imperative versus functional type programming easily based on the situation. Also it doesn't need special data ingestion or indexing pre-processing like Presto. Combining it with Jupyter Notebooks (https://github.com/jupyter-incubator/sparkmagic), one can develop the Spark code in an interactive manner in Scala or Python
Read full review
Return on Investment
Apache
  • Very fast time to market in that so many components are available to use immediately.
  • Error handling mechanisms and patterns of practice are robust and easy to use which in turn has made our application more robust from the start, so fewer bugs.
  • However, testing and debugging routes is more challenging than working is standard Java so that takes more time (less time than writing the components from scratch).
  • Most people don't know Camel coming in and many junior developers find it overwhelming and are not enthusiastic to learn it. So finding people that want to develop/maintain it is a challenge.
Read full review
Apache
  • Faster turn around on feature development, we have seen a noticeable improvement in our agile development since using Spark.
  • Easy adoption, having multiple departments use the same underlying technology even if the use cases are very different allows for more commonality amongst applications which definitely makes the operations team happy.
  • Performance, we have been able to make some applications run over 20x faster since switching to Spark. This has saved us time, headaches, and operating costs.
Read full review
ScreenShots