Dataloader.io delivers a cloud based solution to import and export information from Salesforce.
$99
per month
Informatica PowerCenter (legacy)
Score 9.0 out of 10
N/A
Informatica PowerCenter was data integration technology designed to form the foundation for data integration initiatives, application migration, or analytics. It is a legacy product.
N/A
Pricing
Dataloader.io
Informatica PowerCenter (legacy)
Editions & Modules
Professional
$99.00
per month
Enterprise
$299.00
per month
No answers on this topic
Offerings
Pricing Offerings
Dataloader.io
Informatica PowerCenter (legacy)
Free Trial
No
No
Free/Freemium Version
No
No
Premium Consulting/Integration Services
No
No
Entry-level Setup Fee
No setup fee
No setup fee
Additional Details
—
—
More Pricing Information
Community Pulse
Dataloader.io
Informatica PowerCenter (legacy)
Features
Dataloader.io
Informatica PowerCenter (legacy)
Data Source Connection
Comparison of Data Source Connection features of Product A and Product B
Dataloader.io
-
Ratings
Informatica PowerCenter (legacy)
8.5
18 Ratings
4% above category average
Connect to traditional data sources
00 Ratings
9.018 Ratings
Connecto to Big Data and NoSQL
00 Ratings
8.014 Ratings
Data Transformations
Comparison of Data Transformations features of Product A and Product B
Dataloader.io
-
Ratings
Informatica PowerCenter (legacy)
7.5
18 Ratings
6% below category average
Simple transformations
00 Ratings
8.018 Ratings
Complex transformations
00 Ratings
7.018 Ratings
Data Modeling
Comparison of Data Modeling features of Product A and Product B
Dataloader.io
-
Ratings
Informatica PowerCenter (legacy)
8.2
18 Ratings
5% above category average
Data model creation
00 Ratings
9.015 Ratings
Metadata management
00 Ratings
8.016 Ratings
Business rules and workflow
00 Ratings
9.018 Ratings
Collaboration
00 Ratings
6.116 Ratings
Testing and debugging
00 Ratings
9.017 Ratings
Data Governance
Comparison of Data Governance features of Product A and Product B
Replacing data. If we've put something in a category or a bucket that is no longer named that anymore because we've evolved with the times and we want to rebrand everything, it makes it way easier to do a quick import with the new terms.
1.- Scenaries with poor sources of data is not recomended (Very bad ROI). The solution is for medium-big enterprises with a lot of sources of data and users. 2.- Bank and finance enviroment to integrate differente data form trading, Regulatory reports, decisions makers, fraud and financial crimes because in this kind of scenary the quality of data is the base of the business. 3.- Departments of development and test of applications in enterprises because you can design enviroments, out of the production systems, to development and test the new API's or updateds made.
Extracting Salesforce attachments in original file format! I do not know of a tool that can do this better, or more efficiently! This is a huge benefit to companies that would like to extract attachments from Salesforce for tasks like data migrations.
Cross-object data extract within one file. You can pull data from related objects as long as there is a populated lookup from the object you are extracting, to another object (Child or Parent).
UI is simple and requires very little to no training. Given the acquisition of Mulesoft by Salesforce, I would not be surprised if DataLoader.IO is rolled out as the new global data loading tool for Salesforce.
Informatica Powercenter is an innovative software that works with ETL-type data integration. Connectivity to almost all the database systems.
Great documentation and customer support.
It has a various solution to address data quality issues. data masking, data virtualization. It has various supporting tools or MDM, IDQ, Analyst, BigData which can be used to analyze data and correct it.
At the moment, I can't find a way to rename jobs. This would be useful to organize what was previously created hastily by techs in a rush.
A preview of the job, especially upserts, would take a great deal of stress away from some of us (especially those who are not so confident in their ETL practice).
A native vlookup equivalent may be a welcome addition.
There are too many ways to perform the same or similar functions which in turn makes it challenging to trace what a workflow is doing and at which point (ex. sessions can be designed as static or re-usable and the override can occur at the session or workflow, or both which can be counter productive and confusing when troubleshooting).
The power in structured design is a double edged sword. Simple tasks for a POC can become cumbersome. Ex. if you want to move some data to test a process, you first have to create your sources by importing them which means an ODBC connection or similar will need to be configured, you in turn have to develop your targets and all of the essential building blocks before being able to begin actual development. While I am on sources and targets, I think of a table definition as just that and find it counter intuitive to have to design a table as both a source and target and manage them as different objects. It would be more intuitive to have a table definition and its source/target properties defined by where you drag and drop it in the mapping.
There are no checkpoints or data viewer type functions without designing an entire mapping and workflow. If you would like to simply run a job up to a point and check the throughput, an entire mapping needs to be completed and you would workaround this by creating a flat file target.
It is easy to use and doesn't require a security token, so I enjoy using it. It also doesn't require any download or installation, which is sometimes a blocker to gettingthings done if the company has limits. also, the dataloader.io is easy for other people to pick up, so others can have visibility into the data jobs that have occurred
Dataloader definitely skews towards a more technical userbase. Users should be adept at manipulating data in spreadsheets and decipher JSON formatted error messaging. Additionally, there is a good amount of time need to set up the environment to map to the pertinent fields we are trying to adjust. While I would not recommend the typical account manager to use Dataloader, a typical operations manager should have no issue.
Positives; - Multi User Development Environment - Speed of transformation - Seamless integration between other Informatica products. Negatives; - There should be less windows to maintain developers' focus while using. You probably need 2 big monitors when you start development with Informatica Power Center. - Oracle Analytical functions should be natively used. - E-LT support as well as ETL support.
PowerCenter is robust and fast, and it does a great job meeting all the needs, not just the most commercially vocal needs. In the hands of an expert power user, you can accomplish almost anything with your data. It is not for new users or intermittent users-- for that the Cloud version is a better fit. Be prepared for costly connectors (priced differently for each source or destination you are working with), and just be planful of your projects so you are not paying for connectors you no longer need or want
The utility itself is very self-explanatory and has enough information to guide you through the process. It has an intuitive experience for those familiar with data loading/exporting utilities. Outside of this, they have a Zendesk help center to log support requests and provide documentation to help guide you troubleshoot any issues that may be occurring.
Informatica power center is a leader of the pack of ETL tools and has some great abilities that make it stand out from other ETL tools. It has been a great partner to its clients over a long time so it's definitely dependable. With all the great things about Informatica, it has a bit of tech burden that should be addressed to make it more nimble, reduce the learning curve for new developers, provide better connectivity with visualization tools.
I have used salesforce inspector also for operations like import and export of data from custom objects but it doesn't work well when you have data in huge numbers. Instead of using Salesforce Inspector, one should go for Dataloader.io if the number of records is huge to be dealt with.
While Talend offers a much more comfortable interface to work with, Informatica's forte is performance. And on that front, Informatica Enterprise Data Integration certainly leaves Talend in the dust. For a more back-end-centric use case, Informatica is certainly the ETL tool of choice. On the other hand, if business users would be using the tool, then Talend would be the preferred tool.
HUGE time saving. When we need to clean or review data, we used to have to do it line by line. This can do the work within excel and make cleanup/management an afternoons work as opposed to a week.
Rollback what you did/change/deleted is relatively simple if you remember to back up the data you are manipulating.
The data pipeline automation capability of Informatica means that few resources are needed to pre-process the data that ultimately resides in a Data Warehouse. Once a workflow is implemented, manual intervention is not needed.
PowerCenter did require more resources and time for installation and configuration than was expected/planned for.
The lack of or minimal support of unstructured data means that newer sources of dynamic/changing data cannot be easily processed/transformed through PowerCenter workflows.