Spotfire, formerly known as TIBCO Spotfire, is a visual data science platform that combines visual analytics, data science, and data wrangling, so users can analyze data at-rest and at-scale to solve complex industry-specific problems.
N/A
TIBCO Integration (including BusinessWorks and Flogo)
Score 8.9 out of 10
N/A
TIBCO Integration is used to connect diverse business solutions, data sources, and devices into a single, seamless system.
N/A
Pricing
Spotfire
TIBCO Integration (including BusinessWorks and Flogo)
Editions & Modules
No answers on this topic
No answers on this topic
Offerings
Pricing Offerings
Spotfire
TIBCO Integration (including BusinessWorks and Flogo)
Free Trial
Yes
No
Free/Freemium Version
No
No
Premium Consulting/Integration Services
Yes
Yes
Entry-level Setup Fee
No setup fee
Optional
Additional Details
For Enterprise engagements, contact Spotfire directly for a custom price quote.
—
More Pricing Information
Community Pulse
Spotfire
TIBCO Integration (including BusinessWorks and Flogo)
I don't have too much experience on the alternative tools, but I know Spotfire can perform very well some statistics function in conjunction with TERR. Tableau for example, has more beautiful ways to construct a dashboard and most of time this is why people choose to use it. …
Chose TIBCO Integration (including BusinessWorks and Flogo)
It has given me a good career growth. It's a robust and stable tool with numerous functionalities.
TIBCO has various product suites which can be leveraged by businesses for their different needs like MDM, event based processing, BPEL, Spotfire, and JASPERSOFT.
TIBCO Integration (including BusinessWorks and Flogo)
Likelihood to Recommend
Spotfire
A high level of data integration is available here it supports various data sources and so on. Collaborating features allow users to give access to the dashboard and merge data analytics with other team members. It can meet the demands of both small and large size business enterprises. A customized dashboard and reports are provided to meet the specific needs and get support of extensibility through APIs and customized scripts.
In BusinessWorks it is quite easy to develop applications and monitor where as it is not similar with Scribe. The data which successfully processed in Scribe is quite difficult to view. I think the Scribe console tool should have better view for the data it has processed.At times we founds some of our sales orders are missing which are unable to update/insert in the DB, after investigation found the issue is in Scribe tool. A case has been raised 02129657 for the same and as per the guidelines from the Scribe support engineer we did changes in the shutdown interval to zero. ''update SCRIBEINTERNAL.SCRIBE.KSYNC set keyvalue= 0 where KEYNAME = 'SETTINGS.SHUTDOWNINTERVAL''. Later this change we started getting alert emails like ''Fatal error 325 occurred: Message processor shutting down - defined maximum memory usage (900MB) exceeded. Please contact Scribe support at support@scribesoft.com if this problem persists''. We found this issue is a critical one as the occurrences of it is decreased but still it is persist.
The UI is easy to navigate and map flows are intuitive and easy to build upon after some basic training. However, a user should have some basic knowledge of writing nested statements.
Very responsive customer support.
The debug feature is a great tool for troubleshooting. You can easily identify where the issue is in your flow. At each debug step, it returns the expected value, and when there isn't the expected value, you know when to modify your flow.
The donut chart is I guess a powerful illustrations but I hope it should be done quite simple in Spotfire. But in Spotfire there are lots of steps involve just to build a simple donut chart.
Table calculation (like Row or Column Differences) should be made simple or there should be drag and drop function for Table Calculation. No need for scripting.
Information Link should be changed. If new columns are added to the table just refreshing the data should be able to capture the new column. No need extra step to add column
Missing functionality: We like to run all the files available in the FTP when clicking the run button. Right now, we have to click the run button several times (or have multiple solutions) in order to run files from our FTP site one at a time.
Generating column fields every time a file changes from the source in the FTP is very tedious. Working in integrations makes me have to go back and forth a lot of the time, and doing this is incredibly annoying.
The security is horrendous. We installed an on-premise agent on the customer side, but we don't want them even touching the ETL tool at all. The only knowledge we are fine with knowing and being able to see is that they installed the agent, not the solutions themselves.
Our company's connector would have to point to the dataset name rather than dataset ID. It was very frustrating because we change the dataset name a lot, and then we'd have to repoint it.
It's very difficult to monitor the different integrations that go on because there isn't a consolidated dashboard.
-Easy to distribute information throughout the enterprise using the webplayer. -Ad hoc analysis is possible throughout the enterprise using business author in the webplayer or the thick client. -Low level of support needed by IT team. Access interfaces with LDAP and numerous other authentication methods. -Possible to continually extend the platform with JavaScript, R scripts, HTML, and custom extensions. -Ability to standardize data logic through pre-built queries in the Information Designer. Everyone in the enterprise is using the same logic -Tagging and bookmarking data allows for quick sharing of insights. -Integration with numerous data sources... flat files, data bases, big data, images, etc. -Much improved mapping capability. Also includes the ability to apply data points over any image.
We are deeply entrenched in using Tibco Scribe capabilities, and we are only expanding our usage. It would be one thing if we used it only for a one-time data load, but we have several Scribe maps running constantly, keeping business-critical data up to date. And the ease of use for the occasional, mass data update or upload is simply icing on the cake. I'm a big Microsoft fan, so there is potential down the road to convert our Scribe data integrations to Microsoft Flow (or Power Automate as it is now called). But for now, the functionality just isn't there with Flow (Power Automate), plus the work involved to change all our current integrations would be a large undertaking. So for now and the foreseeable future, Scribe will remain our data integration tool of choice.
Basic tasks like generating meaningful information from large sets of raw data are very easy. The next step of linking to multiple live data sources and linking those tables and performing on the fly analysis of the imported data is understandably more difficult.
As mentioned in the previous sections, setup and maintenance is extremely easy. We don't have many issues for which we need support and there is no need for deep technical skills to use Tibco Cloud Integration platform. The solution provides everything we need for our specific use case, being the replication of our Microsoft Dynamics 365 CRM data to our on premise database for reporting.
Even though, it's a rather stable and predictable tool that's also fast, it does have some bugs and inconsistencies that shut down the system. Depending on the details, it could happen as often as 2-3 times a week, especially during the development period.
We have not had any issues with TIBCO not being available when needed. I have only had to contact support less than 5 times in about 5 year time frame due to syncing issues or a problem with the agent. Support is very quick to respond as well as very helpful.
Generally, the Spotfire client runs with very good performance. There are factors that could affect performance, but normally has to do with loading large analysis files from the library if the database is located some distance away and your global network is not optimal. Once you have your data table(s) loaded in the client application, usually the application is quite good performance-wise.
TIBCO Cloud Integration (including BusinessWorks and Scribe)'s performance of the user interface are not to be complained about. The user interface is swift and is a pleasant user experience. The replication jobs take some time to finish but that is because the number of records to be updated/created on a daily basis is quite big. I did split up the jobs between highrunners (entities with a lot of changes) I update on a daily basis and quite stable entities that I update weekly. That solved my issue of a way to long replication.
Support has been helpful with issues. Support seems to know their product and its capabilities. It would also seem that they have a good sense of the context of the problem; where we are going with this issue and what we want the end outcome to be.
For creating new process, you have lots of palettes to do every develop you need. For created process it is easy to understand even if you have ever seen before it. You can analyze your process in all their detail. User Experience is positive for beginner and expertise people. Just debug mode is still raw, but better in TIBCO BusinessWorks 6.x than TIBCO BusinessWorks 5.x .
The instructor was very in depth and provided relevant training to business users on how to create visualizations. They showed us how to alter settings and filter views, and provided resources for future questions. However, the instructor failed to cover data sources, connecting to data, etc. While it was helpful to see how users can use the data to create reports, they failed to properly instruct us on how to get the dataset in to begin with. We are still trying to figure out connections to certain databases (we have multiple different types).
Before using TIBCO Businessworks, I participated an official training with a specialized instructor. In this training, I learned what I needed to know to build some applications or web services and how to manage its. It was been very important to me. I understood many technical stuff to familiarize with this powerful platform. I can advice TIBCO Educational Services.
The online training is good, provides a good base of knowledge. The video demonstrations were well-done and easy to follow along. Provided exercises are good as well, but I think there could be more challenging exercises. The training has also gone up in price significantly in the last 3 years (in USD, which hurts us even more in Canada), and I'm not sure it is worth the money it now costs (it is worth how much it cost 3 years ago, but not double that.)
The original architecture I created for our implementation had only a particular set of internal business units in mind. Over the years, Spotfire gained in popularity in our company and was being utilized across many more business units. Soon, its usage went beyond what the original architectural implementation could provide. We've since learned about how the product is used by the different teams and are currently in the middle of rolling out a new architecture. I suggest:
Have clearly defined service level agreements with all the teams that will use Spotfire. Your business intelligence group might only need availability during normal working hours, but your production support group might need 24/7 availability. If these groups share one Spotfire server, maintenance of that server might be a problem.
Know the different types of data you will be working with. One group might be working with "public" data while another group might work with sensitive data. Design your Library accordingly and with the proper permissions.
Know the roles of the users of Spotfire. Will there only be a small set of report writers or does everyone have write access to the Library?
ALWAYS add a timestamp prompt to your reports. You don't want multiple users opening a report that will try and pull down millions of rows of data to their local workstations. Another option, of course, is to just hard code a time range in the backing database view (i.e. where activity_date >= sysdate - 90, etc.), but I'd rather educate/train the user base if possible.
This probably goes without saying, but if possible, point to a separate reporting database or a logical standby database. You don't want the company pounding on your primaries and take down your order system.
If we don't design the process correctly, it can do lot of damage (like deleting files or data etc.,) and might assume product issue rather than a design flaw. So it is always recommended to do thorough unit testing , QA and a design review even for a small process to avoid major problems.
Spotfire is significantly ahead of both products from an ETL and data ingestion capability. Spotfire also has substantially better visualizations than Power BI, and although the native visualizations aren't as flexible in Tableau, Spotfire enables users to create completely custom javascript visaualizations, which neither Tableau or Power BI has. Tableau and Power BI are likely only superior to Spotfire with respect to embedded analysis on a website.
Used TIBCO for migration of our data for our learning management system.TIBCO, succesful Content ,data migration(courses, classes, registrations, transcripts etc.) ,User migration(Internal users and Partner migration)and Report migrationIntegrations with different applications Workday,Single Sign-On,CertTracker,Okta,Partner contact sync and Questionmark is succesful.LOD: Learn on demand is a third party vendor application where all Virtual classes are delivered by the vendor application. Learning management system sends the class and registration events to learn on demand via real time web service call through TIBCO
In an enterprise architecture, if Spotfire Advanced Data services(Composite Studio),data marts can be managed optimally and scalability in a data perspective is great. As the web player/consumer is directly proportional to RAM, if the enterprise can handle RAM requirement accomodating fail over mechanisms appropraitely, it is definitely scalable,
We could easily add multiple agents and new systems. This had no impact on the performance. We had some issues because flows cannot be splitted and called by each other. So you have lots of flows which are the same. It would be great if this was improved. Also transferring values via variables between flows seems to be impossible. You can only use global lists with dictionaries.