Azure Synapse Analytics is described as the former Azure SQL Data Warehouse, evolved, and as a limitless analytics service that brings together enterprise data warehousing and Big Data analytics. It gives users the freedom to query data using either serverless or provisioned resources, at scale. Azure Synapse brings these two worlds together with a unified experience to ingest, prepare, manage, and serve data for immediate BI and machine learning needs.
$4,700
per month 5000 Synapse Commit Units (SCUs)
Informatica PowerCenter (legacy)
Score 9.1 out of 10
N/A
Informatica PowerCenter was data integration technology designed to form the foundation for data integration initiatives, application migration, or analytics. It is a legacy product.
N/A
Pricing
Azure Synapse Analytics
Informatica PowerCenter (legacy)
Editions & Modules
Tier 1
$4,700
per month 5,000 Synapse Commit Units (SCUs)
Tier 2
$9,200
per month 10,000 Synapse Commit Units (SCUs)
Tier 3
$21,360
per month 24,000 Synapse Commit Units (SCUs)
Tier 4
$50,400
per month 60,000 Synapse Commit Units (SCUs)
Tier 5
$117,000
per month 150,000 Synapse Commit Units (SCUs)
Tier 6
$259,200
per month 360,000 Synapse Commit Units (SCUs)
No answers on this topic
Offerings
Pricing Offerings
Azure Synapse Analytics
Informatica PowerCenter (legacy)
Free Trial
No
No
Free/Freemium Version
No
No
Premium Consulting/Integration Services
No
No
Entry-level Setup Fee
No setup fee
No setup fee
Additional Details
—
—
More Pricing Information
Community Pulse
Azure Synapse Analytics
Informatica PowerCenter (legacy)
Features
Azure Synapse Analytics
Informatica PowerCenter (legacy)
Data Source Connection
Comparison of Data Source Connection features of Product A and Product B
Azure Synapse Analytics
-
Ratings
Informatica PowerCenter (legacy)
8.5
18 Ratings
3% above category average
Connect to traditional data sources
00 Ratings
9.018 Ratings
Connecto to Big Data and NoSQL
00 Ratings
8.014 Ratings
Data Transformations
Comparison of Data Transformations features of Product A and Product B
Azure Synapse Analytics
-
Ratings
Informatica PowerCenter (legacy)
7.5
18 Ratings
7% below category average
Simple transformations
00 Ratings
8.018 Ratings
Complex transformations
00 Ratings
7.018 Ratings
Data Modeling
Comparison of Data Modeling features of Product A and Product B
Azure Synapse Analytics
-
Ratings
Informatica PowerCenter (legacy)
8.2
18 Ratings
5% above category average
Data model creation
00 Ratings
9.015 Ratings
Metadata management
00 Ratings
8.016 Ratings
Business rules and workflow
00 Ratings
9.018 Ratings
Collaboration
00 Ratings
6.116 Ratings
Testing and debugging
00 Ratings
9.017 Ratings
Data Governance
Comparison of Data Governance features of Product A and Product B
It's well suited for large, fastly growing, and frequently changing data warehouses (e.g., in startups). It's also suited for companies that want a single, relatively easy-to-use, centralized cloud service for all their data needs. Larger, more structured organizations could still benefit from this service by using Synapse Dedicated SQL Pools, knowing that costs will be much higher than other solutions. I think this product is not suited for smaller, simpler workloads (where an Azure SQL Database and a Data Factory could be enough) or very large scenarios, where it may be better to build custom infrastructure.
1.- Scenaries with poor sources of data is not recomended (Very bad ROI). The solution is for medium-big enterprises with a lot of sources of data and users. 2.- Bank and finance enviroment to integrate differente data form trading, Regulatory reports, decisions makers, fraud and financial crimes because in this kind of scenary the quality of data is the base of the business. 3.- Departments of development and test of applications in enterprises because you can design enviroments, out of the production systems, to development and test the new API's or updateds made.
Quick to return data. Queries in a SQL data warehouse architecture tend to return data much more quickly than a OLTP setup. Especially with columnar indexes.
Ability to manage extremely large SQL tables. Our databases contain billions of records. This would be unwieldy without a proper SQL datawarehouse
Backup and replication. Because we're already using SQL, moving the data to a datawarehouse makes it easier to manage as our users are already familiar with SQL.
Informatica Powercenter is an innovative software that works with ETL-type data integration. Connectivity to almost all the database systems.
Great documentation and customer support.
It has a various solution to address data quality issues. data masking, data virtualization. It has various supporting tools or MDM, IDQ, Analyst, BigData which can be used to analyze data and correct it.
With Azure, it's always the same issue, too many moving parts doing similar things with no specialisation. ADF, Fabric Data Factory and Synapse pipeline serve the same purpose. Same goes for Fabric Warehouse and Synapse SQL pools.
Could do better with serverless workloads considering the competition from databricks and its own fabric warehouse
Synapse pipelines is a replica of Azure Data Factory with no tight integration with Synapse and to a surprise, with missing features from ADF. Integration of warehouse can be improved with in environment ETl tools
There are too many ways to perform the same or similar functions which in turn makes it challenging to trace what a workflow is doing and at which point (ex. sessions can be designed as static or re-usable and the override can occur at the session or workflow, or both which can be counter productive and confusing when troubleshooting).
The power in structured design is a double edged sword. Simple tasks for a POC can become cumbersome. Ex. if you want to move some data to test a process, you first have to create your sources by importing them which means an ODBC connection or similar will need to be configured, you in turn have to develop your targets and all of the essential building blocks before being able to begin actual development. While I am on sources and targets, I think of a table definition as just that and find it counter intuitive to have to design a table as both a source and target and manage them as different objects. It would be more intuitive to have a table definition and its source/target properties defined by where you drag and drop it in the mapping.
There are no checkpoints or data viewer type functions without designing an entire mapping and workflow. If you would like to simply run a job up to a point and check the throughput, an entire mapping needs to be completed and you would workaround this by creating a flat file target.
The data warehouse portion is very much like old style on-prem SQL server, so most SQL skills one has mastered carry over easily. Azure Data Factory has an easy drag and drop system which allows quick building of pipelines with minimal coding. The Spark portion is the only really complex portion, but if there's an in-house python expert, then the Spark portion is also quiet useable.
Positives; - Multi User Development Environment - Speed of transformation - Seamless integration between other Informatica products. Negatives; - There should be less windows to maintain developers' focus while using. You probably need 2 big monitors when you start development with Informatica Power Center. - Oracle Analytical functions should be natively used. - E-LT support as well as ETL support.
PowerCenter is robust and fast, and it does a great job meeting all the needs, not just the most commercially vocal needs. In the hands of an expert power user, you can accomplish almost anything with your data. It is not for new users or intermittent users-- for that the Cloud version is a better fit. Be prepared for costly connectors (priced differently for each source or destination you are working with), and just be planful of your projects so you are not paying for connectors you no longer need or want
Microsoft does its best to support Synapse. More and more articles are being added to the documentation, providing more useful information on best utilizing its features. The examples provided work well for basic knowledge, but more complex examples should be added to further assist in discovering the vast abilities that the system has.
Informatica power center is a leader of the pack of ETL tools and has some great abilities that make it stand out from other ETL tools. It has been a great partner to its clients over a long time so it's definitely dependable. With all the great things about Informatica, it has a bit of tech burden that should be addressed to make it more nimble, reduce the learning curve for new developers, provide better connectivity with visualization tools.
In comparing Azure Synapse to the Google BigQuery - the biggest highlight that I'd like to bring forward is Azure Synapse SQL leverages a scale-out architecture in order to distribute computational processing of data across multiple nodes whereas Google BigQuery only takes into account computation and storage.
While Talend offers a much more comfortable interface to work with, Informatica's forte is performance. And on that front, Informatica Enterprise Data Integration certainly leaves Talend in the dust. For a more back-end-centric use case, Informatica is certainly the ETL tool of choice. On the other hand, if business users would be using the tool, then Talend would be the preferred tool.
Licensing fees is replaced with Azure subscription fee. No big saving there
More visibility into the Azure usage and cost
It can be used a hot storage and old data can be archived to data lake. Real time data integration is possible via external tables and Microsoft Power BI
The data pipeline automation capability of Informatica means that few resources are needed to pre-process the data that ultimately resides in a Data Warehouse. Once a workflow is implemented, manual intervention is not needed.
PowerCenter did require more resources and time for installation and configuration than was expected/planned for.
The lack of or minimal support of unstructured data means that newer sources of dynamic/changing data cannot be easily processed/transformed through PowerCenter workflows.