TIBCO Integration (including BusinessWorks and Flogo)
Likelihood to Recommend
Red Hat
JBoss EAP is subscription based/open source platform. It's very reliable and great for deploying high transaction Java based enterprise applications. It integrates well with third party components like mod_cluster and supports popular Java EE web-based frameworks such as Spring, Angular JS, jQuery Mobile, and Google Web Toolkit.
In BusinessWorks it is quite easy to develop applications and monitor where as it is not similar with Scribe. The data which successfully processed in Scribe is quite difficult to view. I think the Scribe console tool should have better view for the data it has processed.At times we founds some of our sales orders are missing which are unable to update/insert in the DB, after investigation found the issue is in Scribe tool. A case has been raised 02129657 for the same and as per the guidelines from the Scribe support engineer we did changes in the shutdown interval to zero. ''update SCRIBEINTERNAL.SCRIBE.KSYNC set keyvalue= 0 where KEYNAME = 'SETTINGS.SHUTDOWNINTERVAL''. Later this change we started getting alert emails like ''Fatal error 325 occurred: Message processor shutting down - defined maximum memory usage (900MB) exceeded. Please contact Scribe support at support@scribesoft.com if this problem persists''. We found this issue is a critical one as the occurrences of it is decreased but still it is persist.
MOD_CLUSTER integration. JBoss EAP integrates pretty well with mod_cluster. This is an intelligent load balancer especially useful in highly clustered environments.
Supports enterprise-grade features such as high availability clustering, distributed caching, messaging etc.
Supports deployment in on-premise, virtual and hybrid cloud environments.
The UI is easy to navigate and map flows are intuitive and easy to build upon after some basic training. However, a user should have some basic knowledge of writing nested statements.
Very responsive customer support.
The debug feature is a great tool for troubleshooting. You can easily identify where the issue is in your flow. At each debug step, it returns the expected value, and when there isn't the expected value, you know when to modify your flow.
Jboss CLI is a great tool but we had trouble using it to get values that are displayed on Jboss GUI. It also has limitations parsing the applications.xml files and we had to use a mix of jboss-cli and linux bash commands to automate certain application administrative tasks.
JBoss doesn't really provides performance tuning recommendations. It would have been nice if it could learn from the current demand vs current settings for things like connection pool, server configurations, garbage collection etc.
Missing functionality: We like to run all the files available in the FTP when clicking the run button. Right now, we have to click the run button several times (or have multiple solutions) in order to run files from our FTP site one at a time.
Generating column fields every time a file changes from the source in the FTP is very tedious. Working in integrations makes me have to go back and forth a lot of the time, and doing this is incredibly annoying.
The security is horrendous. We installed an on-premise agent on the customer side, but we don't want them even touching the ETL tool at all. The only knowledge we are fine with knowing and being able to see is that they installed the agent, not the solutions themselves.
Our company's connector would have to point to the dataset name rather than dataset ID. It was very frustrating because we change the dataset name a lot, and then we'd have to repoint it.
It's very difficult to monitor the different integrations that go on because there isn't a consolidated dashboard.
We are deeply entrenched in using Tibco Scribe capabilities, and we are only expanding our usage. It would be one thing if we used it only for a one-time data load, but we have several Scribe maps running constantly, keeping business-critical data up to date. And the ease of use for the occasional, mass data update or upload is simply icing on the cake. I'm a big Microsoft fan, so there is potential down the road to convert our Scribe data integrations to Microsoft Flow (or Power Automate as it is now called). But for now, the functionality just isn't there with Flow (Power Automate), plus the work involved to change all our current integrations would be a large undertaking. So for now and the foreseeable future, Scribe will remain our data integration tool of choice.
As mentioned in the previous sections, setup and maintenance is extremely easy. We don't have many issues for which we need support and there is no need for deep technical skills to use Tibco Cloud Integration platform. The solution provides everything we need for our specific use case, being the replication of our Microsoft Dynamics 365 CRM data to our on premise database for reporting.
We have not had any issues with TIBCO not being available when needed. I have only had to contact support less than 5 times in about 5 year time frame due to syncing issues or a problem with the agent. Support is very quick to respond as well as very helpful.
Usually, Red Hat JBoss Enterprise Application Platform is good at performance and well suited for high traffic Java EE-based applications, but we have faced hard times performance tuning it for our specific needs. The product would be nicer if they would add a performance diagnostic and recommendations feature to it.
TIBCO Cloud Integration (including BusinessWorks and Scribe)'s performance of the user interface are not to be complained about. The user interface is swift and is a pleasant user experience. The replication jobs take some time to finish but that is because the number of records to be updated/created on a daily basis is quite big. I did split up the jobs between highrunners (entities with a lot of changes) I update on a daily basis and quite stable entities that I update weekly. That solved my issue of a way to long replication.
For creating new process, you have lots of palettes to do every develop you need. For created process it is easy to understand even if you have ever seen before it. You can analyze your process in all their detail. User Experience is positive for beginner and expertise people. Just debug mode is still raw, but better in TIBCO BusinessWorks 6.x than TIBCO BusinessWorks 5.x .
Before using TIBCO Businessworks, I participated an official training with a specialized instructor. In this training, I learned what I needed to know to build some applications or web services and how to manage its. It was been very important to me. I understood many technical stuff to familiarize with this powerful platform. I can advice TIBCO Educational Services.
If we don't design the process correctly, it can do lot of damage (like deleting files or data etc.,) and might assume product issue rather than a design flaw. So it is always recommended to do thorough unit testing , QA and a design review even for a small process to avoid major problems.
Used TIBCO for migration of our data for our learning management system.TIBCO, succesful Content ,data migration(courses, classes, registrations, transcripts etc.) ,User migration(Internal users and Partner migration)and Report migrationIntegrations with different applications Workday,Single Sign-On,CertTracker,Okta,Partner contact sync and Questionmark is succesful.LOD: Learn on demand is a third party vendor application where all Virtual classes are delivered by the vendor application. Learning management system sends the class and registration events to learn on demand via real time web service call through TIBCO
We could easily add multiple agents and new systems. This had no impact on the performance. We had some issues because flows cannot be splitted and called by each other. So you have lots of flows which are the same. It would be great if this was improved. Also transferring values via variables between flows seems to be impossible. You can only use global lists with dictionaries.