Informatica PowerCenter was data integration technology designed to form the foundation for data integration initiatives, application migration, or analytics. It is a legacy product.
N/A
Spotfire
Score 8.2 out of 10
N/A
Spotfire, formerly known as TIBCO Spotfire, is a visual data science platform that combines visual analytics, data science, and data wrangling, so users can analyze data at-rest and at-scale to solve complex industry-specific problems.
N/A
Pricing
Informatica PowerCenter (legacy)
Spotfire
Editions & Modules
No answers on this topic
No answers on this topic
Offerings
Pricing Offerings
Informatica PowerCenter (legacy)
Spotfire
Free Trial
No
Yes
Free/Freemium Version
No
No
Premium Consulting/Integration Services
No
Yes
Entry-level Setup Fee
No setup fee
No setup fee
Additional Details
—
For Enterprise engagements, contact Spotfire directly for a custom price quote.
More Pricing Information
Community Pulse
Informatica PowerCenter (legacy)
Spotfire
Features
Informatica PowerCenter (legacy)
Spotfire
Data Source Connection
Comparison of Data Source Connection features of Product A and Product B
Informatica PowerCenter (legacy)
8.5
18 Ratings
3% above category average
Spotfire
-
Ratings
Connect to traditional data sources
9.018 Ratings
00 Ratings
Connecto to Big Data and NoSQL
8.014 Ratings
00 Ratings
Data Transformations
Comparison of Data Transformations features of Product A and Product B
Informatica PowerCenter (legacy)
7.5
18 Ratings
8% below category average
Spotfire
-
Ratings
Simple transformations
8.018 Ratings
00 Ratings
Complex transformations
7.018 Ratings
00 Ratings
Data Modeling
Comparison of Data Modeling features of Product A and Product B
Informatica PowerCenter (legacy)
8.2
18 Ratings
5% above category average
Spotfire
-
Ratings
Data model creation
9.015 Ratings
00 Ratings
Metadata management
8.016 Ratings
00 Ratings
Business rules and workflow
9.018 Ratings
00 Ratings
Collaboration
6.116 Ratings
00 Ratings
Testing and debugging
9.017 Ratings
00 Ratings
Data Governance
Comparison of Data Governance features of Product A and Product B
Informatica PowerCenter (legacy)
9.0
15 Ratings
12% above category average
Spotfire
-
Ratings
Integration with data quality tools
9.015 Ratings
00 Ratings
Integration with MDM tools
9.013 Ratings
00 Ratings
Platform Connectivity
Comparison of Platform Connectivity features of Product A and Product B
Informatica PowerCenter (legacy)
-
Ratings
Spotfire
7.2
8 Ratings
15% below category average
Connect to Multiple Data Sources
00 Ratings
7.88 Ratings
Extend Existing Data Sources
00 Ratings
7.48 Ratings
Automatic Data Format Detection
00 Ratings
7.88 Ratings
MDM Integration
00 Ratings
6.05 Ratings
Data Exploration
Comparison of Data Exploration features of Product A and Product B
Informatica PowerCenter (legacy)
-
Ratings
Spotfire
9.1
8 Ratings
7% above category average
Visualization
00 Ratings
9.08 Ratings
Interactive Data Analysis
00 Ratings
9.28 Ratings
Data Preparation
Comparison of Data Preparation features of Product A and Product B
Informatica PowerCenter (legacy)
-
Ratings
Spotfire
7.4
8 Ratings
10% below category average
Interactive Data Cleaning and Enrichment
00 Ratings
7.28 Ratings
Data Transformations
00 Ratings
8.08 Ratings
Data Encryption
00 Ratings
7.05 Ratings
Built-in Processors
00 Ratings
7.55 Ratings
Platform Data Modeling
Comparison of Platform Data Modeling features of Product A and Product B
Informatica PowerCenter (legacy)
-
Ratings
Spotfire
7.6
8 Ratings
10% below category average
Multiple Model Development Languages and Tools
00 Ratings
7.57 Ratings
Automated Machine Learning
00 Ratings
8.55 Ratings
Single platform for multiple model development
00 Ratings
7.68 Ratings
Self-Service Model Delivery
00 Ratings
6.76 Ratings
Model Deployment
Comparison of Model Deployment features of Product A and Product B
1.- Scenaries with poor sources of data is not recomended (Very bad ROI). The solution is for medium-big enterprises with a lot of sources of data and users. 2.- Bank and finance enviroment to integrate differente data form trading, Regulatory reports, decisions makers, fraud and financial crimes because in this kind of scenary the quality of data is the base of the business. 3.- Departments of development and test of applications in enterprises because you can design enviroments, out of the production systems, to development and test the new API's or updateds made.
A high level of data integration is available here it supports various data sources and so on. Collaborating features allow users to give access to the dashboard and merge data analytics with other team members. It can meet the demands of both small and large size business enterprises. A customized dashboard and reports are provided to meet the specific needs and get support of extensibility through APIs and customized scripts.
Informatica Powercenter is an innovative software that works with ETL-type data integration. Connectivity to almost all the database systems.
Great documentation and customer support.
It has a various solution to address data quality issues. data masking, data virtualization. It has various supporting tools or MDM, IDQ, Analyst, BigData which can be used to analyze data and correct it.
There are too many ways to perform the same or similar functions which in turn makes it challenging to trace what a workflow is doing and at which point (ex. sessions can be designed as static or re-usable and the override can occur at the session or workflow, or both which can be counter productive and confusing when troubleshooting).
The power in structured design is a double edged sword. Simple tasks for a POC can become cumbersome. Ex. if you want to move some data to test a process, you first have to create your sources by importing them which means an ODBC connection or similar will need to be configured, you in turn have to develop your targets and all of the essential building blocks before being able to begin actual development. While I am on sources and targets, I think of a table definition as just that and find it counter intuitive to have to design a table as both a source and target and manage them as different objects. It would be more intuitive to have a table definition and its source/target properties defined by where you drag and drop it in the mapping.
There are no checkpoints or data viewer type functions without designing an entire mapping and workflow. If you would like to simply run a job up to a point and check the throughput, an entire mapping needs to be completed and you would workaround this by creating a flat file target.
The donut chart is I guess a powerful illustrations but I hope it should be done quite simple in Spotfire. But in Spotfire there are lots of steps involve just to build a simple donut chart.
Table calculation (like Row or Column Differences) should be made simple or there should be drag and drop function for Table Calculation. No need for scripting.
Information Link should be changed. If new columns are added to the table just refreshing the data should be able to capture the new column. No need extra step to add column
-Easy to distribute information throughout the enterprise using the webplayer. -Ad hoc analysis is possible throughout the enterprise using business author in the webplayer or the thick client. -Low level of support needed by IT team. Access interfaces with LDAP and numerous other authentication methods. -Possible to continually extend the platform with JavaScript, R scripts, HTML, and custom extensions. -Ability to standardize data logic through pre-built queries in the Information Designer. Everyone in the enterprise is using the same logic -Tagging and bookmarking data allows for quick sharing of insights. -Integration with numerous data sources... flat files, data bases, big data, images, etc. -Much improved mapping capability. Also includes the ability to apply data points over any image.
Positives; - Multi User Development Environment - Speed of transformation - Seamless integration between other Informatica products. Negatives; - There should be less windows to maintain developers' focus while using. You probably need 2 big monitors when you start development with Informatica Power Center. - Oracle Analytical functions should be natively used. - E-LT support as well as ETL support.
Basic tasks like generating meaningful information from large sets of raw data are very easy. The next step of linking to multiple live data sources and linking those tables and performing on the fly analysis of the imported data is understandably more difficult.
Even though, it's a rather stable and predictable tool that's also fast, it does have some bugs and inconsistencies that shut down the system. Depending on the details, it could happen as often as 2-3 times a week, especially during the development period.
PowerCenter is robust and fast, and it does a great job meeting all the needs, not just the most commercially vocal needs. In the hands of an expert power user, you can accomplish almost anything with your data. It is not for new users or intermittent users-- for that the Cloud version is a better fit. Be prepared for costly connectors (priced differently for each source or destination you are working with), and just be planful of your projects so you are not paying for connectors you no longer need or want
Generally, the Spotfire client runs with very good performance. There are factors that could affect performance, but normally has to do with loading large analysis files from the library if the database is located some distance away and your global network is not optimal. Once you have your data table(s) loaded in the client application, usually the application is quite good performance-wise.
Informatica power center is a leader of the pack of ETL tools and has some great abilities that make it stand out from other ETL tools. It has been a great partner to its clients over a long time so it's definitely dependable. With all the great things about Informatica, it has a bit of tech burden that should be addressed to make it more nimble, reduce the learning curve for new developers, provide better connectivity with visualization tools.
Support has been helpful with issues. Support seems to know their product and its capabilities. It would also seem that they have a good sense of the context of the problem; where we are going with this issue and what we want the end outcome to be.
The instructor was very in depth and provided relevant training to business users on how to create visualizations. They showed us how to alter settings and filter views, and provided resources for future questions. However, the instructor failed to cover data sources, connecting to data, etc. While it was helpful to see how users can use the data to create reports, they failed to properly instruct us on how to get the dataset in to begin with. We are still trying to figure out connections to certain databases (we have multiple different types).
The online training is good, provides a good base of knowledge. The video demonstrations were well-done and easy to follow along. Provided exercises are good as well, but I think there could be more challenging exercises. The training has also gone up in price significantly in the last 3 years (in USD, which hurts us even more in Canada), and I'm not sure it is worth the money it now costs (it is worth how much it cost 3 years ago, but not double that.)
The original architecture I created for our implementation had only a particular set of internal business units in mind. Over the years, Spotfire gained in popularity in our company and was being utilized across many more business units. Soon, its usage went beyond what the original architectural implementation could provide. We've since learned about how the product is used by the different teams and are currently in the middle of rolling out a new architecture. I suggest:
Have clearly defined service level agreements with all the teams that will use Spotfire. Your business intelligence group might only need availability during normal working hours, but your production support group might need 24/7 availability. If these groups share one Spotfire server, maintenance of that server might be a problem.
Know the different types of data you will be working with. One group might be working with "public" data while another group might work with sensitive data. Design your Library accordingly and with the proper permissions.
Know the roles of the users of Spotfire. Will there only be a small set of report writers or does everyone have write access to the Library?
ALWAYS add a timestamp prompt to your reports. You don't want multiple users opening a report that will try and pull down millions of rows of data to their local workstations. Another option, of course, is to just hard code a time range in the backing database view (i.e. where activity_date >= sysdate - 90, etc.), but I'd rather educate/train the user base if possible.
This probably goes without saying, but if possible, point to a separate reporting database or a logical standby database. You don't want the company pounding on your primaries and take down your order system.
While Talend offers a much more comfortable interface to work with, Informatica's forte is performance. And on that front, Informatica Enterprise Data Integration certainly leaves Talend in the dust. For a more back-end-centric use case, Informatica is certainly the ETL tool of choice. On the other hand, if business users would be using the tool, then Talend would be the preferred tool.
Spotfire is significantly ahead of both products from an ETL and data ingestion capability. Spotfire also has substantially better visualizations than Power BI, and although the native visualizations aren't as flexible in Tableau, Spotfire enables users to create completely custom javascript visaualizations, which neither Tableau or Power BI has. Tableau and Power BI are likely only superior to Spotfire with respect to embedded analysis on a website.
In an enterprise architecture, if Spotfire Advanced Data services(Composite Studio),data marts can be managed optimally and scalability in a data perspective is great. As the web player/consumer is directly proportional to RAM, if the enterprise can handle RAM requirement accomodating fail over mechanisms appropraitely, it is definitely scalable,
The data pipeline automation capability of Informatica means that few resources are needed to pre-process the data that ultimately resides in a Data Warehouse. Once a workflow is implemented, manual intervention is not needed.
PowerCenter did require more resources and time for installation and configuration than was expected/planned for.
The lack of or minimal support of unstructured data means that newer sources of dynamic/changing data cannot be easily processed/transformed through PowerCenter workflows.