Azuqua was a tool that helped users integrate their SaaS applications and build custom automations. It was acquired by Okta in late 2019, and is now part of Okta Workflows. Okta Workflows leverages Azuqua’s workflow orchestration engine and application integrations to automate complex identity-centric processes such as user onboarding and offboarding. The product is available as part of the Okta Lifecycle Management…
N/A
TIBCO Integration (including BusinessWorks and Flogo)
Score 8.9 out of 10
N/A
TIBCO Integration is used to connect diverse business solutions, data sources, and devices into a single, seamless system.
N/A
Pricing
Okta Workflows
TIBCO Integration (including BusinessWorks and Flogo)
Editions & Modules
No answers on this topic
No answers on this topic
Offerings
Pricing Offerings
Okta Workflows
TIBCO Integration (including BusinessWorks and Flogo)
Free Trial
No
No
Free/Freemium Version
No
No
Premium Consulting/Integration Services
No
Yes
Entry-level Setup Fee
Optional
Optional
Additional Details
—
—
More Pricing Information
Community Pulse
Okta Workflows
TIBCO Integration (including BusinessWorks and Flogo)
Features
Okta Workflows
TIBCO Integration (including BusinessWorks and Flogo)
Cloud Data Integration
Comparison of Cloud Data Integration features of Product A and Product B
Okta Workflows
-
Ratings
TIBCO Integration (including BusinessWorks and Flogo)
5.8
128 Ratings
32% below category average
Pre-built connectors
00 Ratings
7.0121 Ratings
Connector modification
00 Ratings
5.58 Ratings
Support for real-time and batch integration
00 Ratings
6.4117 Ratings
Data quality services
00 Ratings
5.56 Ratings
Data security features
00 Ratings
2.77 Ratings
Monitoring console
00 Ratings
7.8121 Ratings
Best Alternatives
Okta Workflows
TIBCO Integration (including BusinessWorks and Flogo)
TIBCO Integration (including BusinessWorks and Flogo)
Likelihood to Recommend
Okta
Azuqua is well suited to connect data based systems or to add an extra level of automation to Smartsheet without requiring the control center. It is also well suited for people who don't have in depth understandings of programming. The UI is mostly visual with click and drag systems instead of requiring manually entered variables.
In BusinessWorks it is quite easy to develop applications and monitor where as it is not similar with Scribe. The data which successfully processed in Scribe is quite difficult to view. I think the Scribe console tool should have better view for the data it has processed.At times we founds some of our sales orders are missing which are unable to update/insert in the DB, after investigation found the issue is in Scribe tool. A case has been raised 02129657 for the same and as per the guidelines from the Scribe support engineer we did changes in the shutdown interval to zero. ''update SCRIBEINTERNAL.SCRIBE.KSYNC set keyvalue= 0 where KEYNAME = 'SETTINGS.SHUTDOWNINTERVAL''. Later this change we started getting alert emails like ''Fatal error 325 occurred: Message processor shutting down - defined maximum memory usage (900MB) exceeded. Please contact Scribe support at support@scribesoft.com if this problem persists''. We found this issue is a critical one as the occurrences of it is decreased but still it is persist.
The concept of reduced code to simplify use by less technical teams lowers the barriers to integration and allows teams to collaborate with ideas and concepts much easier
The ability to review simply any error cases simplifies the old approaches of debugging and reviewing large and complex logs
While not strictly part of the platform the support team's efforts to assist, to help clarify issues and then (where necessary) to resolve bugs was a large benefit and a key driver to extend the platform's footprint.
The UI is easy to navigate and map flows are intuitive and easy to build upon after some basic training. However, a user should have some basic knowledge of writing nested statements.
Very responsive customer support.
The debug feature is a great tool for troubleshooting. You can easily identify where the issue is in your flow. At each debug step, it returns the expected value, and when there isn't the expected value, you know when to modify your flow.
The lack of connection/card documentation. Every card does have a section with details, but they are sometimes lacking.
The help center and community also need some structuring work. Every single connection/app should have a section with detailed documentation regarding its triggers and actions.
The FLO history section needs to be more refined. It sometimes does not load and choosing the date doesn't actually show execution results from that particular day.
Missing functionality: We like to run all the files available in the FTP when clicking the run button. Right now, we have to click the run button several times (or have multiple solutions) in order to run files from our FTP site one at a time.
Generating column fields every time a file changes from the source in the FTP is very tedious. Working in integrations makes me have to go back and forth a lot of the time, and doing this is incredibly annoying.
The security is horrendous. We installed an on-premise agent on the customer side, but we don't want them even touching the ETL tool at all. The only knowledge we are fine with knowing and being able to see is that they installed the agent, not the solutions themselves.
Our company's connector would have to point to the dataset name rather than dataset ID. It was very frustrating because we change the dataset name a lot, and then we'd have to repoint it.
It's very difficult to monitor the different integrations that go on because there isn't a consolidated dashboard.
We are deeply entrenched in using Tibco Scribe capabilities, and we are only expanding our usage. It would be one thing if we used it only for a one-time data load, but we have several Scribe maps running constantly, keeping business-critical data up to date. And the ease of use for the occasional, mass data update or upload is simply icing on the cake. I'm a big Microsoft fan, so there is potential down the road to convert our Scribe data integrations to Microsoft Flow (or Power Automate as it is now called). But for now, the functionality just isn't there with Flow (Power Automate), plus the work involved to change all our current integrations would be a large undertaking. So for now and the foreseeable future, Scribe will remain our data integration tool of choice.
The system is working as it should, keeping our programs safe from outside hackers. Helping us keep our passwords safe, convenient and already ready to get us logged into the program securely and quickly. Verification that only authorized users are able to access our company's programs. Okta Workflows (Azuqua) is a very good system that has helped our company greatly.
As mentioned in the previous sections, setup and maintenance is extremely easy. We don't have many issues for which we need support and there is no need for deep technical skills to use Tibco Cloud Integration platform. The solution provides everything we need for our specific use case, being the replication of our Microsoft Dynamics 365 CRM data to our on premise database for reporting.
We have not had any issues with TIBCO not being available when needed. I have only had to contact support less than 5 times in about 5 year time frame due to syncing issues or a problem with the agent. Support is very quick to respond as well as very helpful.
TIBCO Cloud Integration (including BusinessWorks and Scribe)'s performance of the user interface are not to be complained about. The user interface is swift and is a pleasant user experience. The replication jobs take some time to finish but that is because the number of records to be updated/created on a daily basis is quite big. I did split up the jobs between highrunners (entities with a lot of changes) I update on a daily basis and quite stable entities that I update weekly. That solved my issue of a way to long replication.
For creating new process, you have lots of palettes to do every develop you need. For created process it is easy to understand even if you have ever seen before it. You can analyze your process in all their detail. User Experience is positive for beginner and expertise people. Just debug mode is still raw, but better in TIBCO BusinessWorks 6.x than TIBCO BusinessWorks 5.x .
Before using TIBCO Businessworks, I participated an official training with a specialized instructor. In this training, I learned what I needed to know to build some applications or web services and how to manage its. It was been very important to me. I understood many technical stuff to familiarize with this powerful platform. I can advice TIBCO Educational Services.
If we don't design the process correctly, it can do lot of damage (like deleting files or data etc.,) and might assume product issue rather than a design flaw. So it is always recommended to do thorough unit testing , QA and a design review even for a small process to avoid major problems.
I had to use the Automate tool for funneling image assets in bulk (tens of thousands) from FTPs into various destinations on an eCommerce platform. The user interface was quite harsh in comparison to Azuqua. Far more text/code line driven.
Used TIBCO for migration of our data for our learning management system.TIBCO, succesful Content ,data migration(courses, classes, registrations, transcripts etc.) ,User migration(Internal users and Partner migration)and Report migrationIntegrations with different applications Workday,Single Sign-On,CertTracker,Okta,Partner contact sync and Questionmark is succesful.LOD: Learn on demand is a third party vendor application where all Virtual classes are delivered by the vendor application. Learning management system sends the class and registration events to learn on demand via real time web service call through TIBCO
We could easily add multiple agents and new systems. This had no impact on the performance. We had some issues because flows cannot be splitted and called by each other. So you have lots of flows which are the same. It would be great if this was improved. Also transferring values via variables between flows seems to be impossible. You can only use global lists with dictionaries.