Benchling’s cloud platform is used by pharmaceutical and biotech companies to analyze complex datasets, streamline research workflows, design DNA using CRISPR gene-editing technology and more.
$0
per month
Nintex
Score 8.0 out of 10
N/A
Nintex offers a platform that helps companies discover, automate, and optimize business processes.
$480
Minimum 1,000 users per user
Teradata Vantage
Score 8.1 out of 10
N/A
Teradata Vantage is presented as a modern analytics cloud platform that unifies everything—data lakes, data warehouses, analytics, and new data sources and types. Supports hybrid multi-cloud environments and priced for flexibility, Vantage delivers unlimited intelligence to build the future of business.
Users can deploy Vantage on public clouds (such as AWS, Azure, and GCP), hybrid multi-cloud environments, on-premises with Teradata IntelliFlex, or on commodity hardware with VMware.
Benchling is especially well suited to groups or contexts where there are many users who do not have a coding background but need a seamless and structured approach to data. Benchling is particularly useful in cases where there are data flows from instruments and other devices where the data can be deposited in an automated fashion. It is likely less appropriate or useful to users who are just looking for a general data warehouse solution.
We use Nintex to automate fundraising outreach at scale. It helps us send personalized emails to a large contact list, and we’d also like to automate follow-ups when there’s no reply. If you need highly customized solutions or clean, fully controllable code, I wouldn’t recommend Nintex. It has many features, but it’s not the same as building your own system from scratch. That said, it can save a lot of time for standard automation workflows.
Teradata Vantage is well suited for large scale ETL pipelines like the ones we developed for anti money laundering risk matrices. It handles heavy joins, aggregations, and transformations on transactional data efficiently. We generate alert variables, adjust for inflation, and monitor establishments monthly with it, all integrated with Python and Control-M for a centralised automation across the company. For less appropriate, I would say that heavy resource demands might slow down experimentation for iterative work.
Integrations with other services using various secure authentication methods, along with the seamless integration with SharePoint, are the icing on the cake. This makes it superior to other BPM tools available in the market.
Flexibility in application development - The diverse configurable properties offer multiple ways to utilise the controls and events, affording the flexibility to expand your scope and enabling the creation and use of processes in a myriad of ways.
The streamlined and efficient deployment process significantly accelerates release management, allowing for faster and smoother implementation of updates and new features.
The user interface of the pages offers a more refined and appealing look and feel compared to most other BPM tools.
Some of the integrations can be a bit spotty so it depends on what kind of data source you are integrating
Sometimes new users are not always aware of all the various functionality that Benchling has - can do better to provide more user awareness of more complex features
If you are creating a process with parallel subprocesses, there's no way to see, in a single view in Nintex, all the steps for the subprocesses. You have to view each sub-process in its own view, so it's hard to see what's going on at a high level.
There isn't an easy way to filter the processes by another user (not yourself) in Nintex. There is a report that shows processes and objects by user, but that's not as convenient. This is something that I've seen in other tools (OpenPages by IBM) so I am surprised that it is missing.
Nintex doesn't really have a way to capture iterative processes (which we have a lot of). It's designed for linear processes.
Teradata is an excellent option but only for a massive amount of data warehousing or analysis. If your data is not that big then it could be a misfit for your company and cost you a lot. The cost associated is quite extensive as compared to some other alternative RDBMS systems available in the market.
Migration of data from Teradata to some other RDBMS systems is quite painful as the transition is not that smooth and you need to follow many steps and even if one of them fails. You need to start from the beginning almost.
Last but not least the UI is pretty outdated and needs a revamp. Though it is simple, it needs to be presented in a much better way and more advanced options need to bee presented on the front page itself.
We are currently investigating which collaboration platform best suits our needs. Chances are that we move to SharePoint Online and then we're going to also consider the microsoft power platform (power automate and power apps) to develop forms and workflows. Aspecially the pricing model for the cloud is currently a blocking factor to go for the Nintex solution in the Cloud.
Teradata is a mature RDBMS system that expands its functionality towards the current cloud capabilities like object storage and flexible compute scale.
Based on the on-prem experience with this tool, I believe that they have a lot of potential to help the online version catch up to where the on-prem left off. Nintex developed their online version and it is not as fully formed or capable compared to the on-prem version, and the licensing model scales back what we would have liked to be an expansion or at least continuous improvement of existing flows. It is also not near as user friendly specifically to non-developers and has an uncanny similarity to Microsoft Flow in the online instance. Consistent with my reviews of the tool - I believe they have some good approaches to design thinking that, if translated well from on-prem to online, could make this a clear winner again.
Teradata Vantage allows us to create a scalable infrastructure to support our strategic initiatives. The dedicated compute power ensures reliable performance with isolated workloads and dedicated resources, optimizing workflows for faster, more efficient data transfers. The compute clusters support ETL processes and OSF’s developers and data science team with the flexibility to create self-service analytics, to spin up/down at any time, driving better performance and minimizing costs.
The Nintex Process Platform has never crashed or had any availability issues during my usage. However there was an issue that was of my own making that caused a slowdown of the system. I had set up a process to run once a day and check for employees on a list that had certain parameters selected, and for some reason that I had to troubleshoot, the process instead ran constantly, which filled the cache quickly. I ended up having to dismantle that process so the system didn't crash.
Unlike any other process automation product out there. Not only is it a low-code, easy to use tool for building processes in environments like SharePoint or Salesforce, they have really started to expand their tool-set by offering tools to manage other things like process mapping, RPA, mobile,etc.
The support team works as fast as they can and they are usually fast to solver the issues. Sometimes they need more time to solve one of them because our workflows and so on are more complex than usual clients.
We have meetings at the beginning with the technical team to explain our requirements to them and they were really putting in a lot of effort to come up with a solution which will address all our needs. They implemented the software and also trained a few of our resources on the same too. We can get in touch with them now as well whenever we run into a roadblock but it's very less now.
I used the Nintex training software, it was easy to watch and follow along. It didn't go too fast and was descriptive enough to understand what the steps needed were in order to produce efficient workflows and user friendly forms.
1.Start with Simple Workflows: Begin with basic workflows to gain user confidence before tackling complex processes. 2.Involve Stakeholders Early: Engage business users and IT early to align workflows with real business needs. 3.Comprehensive Training: Invest in user training to ensure smooth adoption and reduce resistance. 4.Leverage Prebuilt Templates: Use Nintex’s templates to speed up implementation and maintain consistency. 5.Iterate and Optimize: Continuously improve workflows based on user feedback and performance metrics.
Benchling was much more of a full stack solution and provide much more features that were relevant to the group. Airtable was more of a generic way to manage large amounts of data, but the complexity was still high for the types of data that would be need to be managed and there would need to be some workarounds. Overall Benchling was selected since it also had an electronic lab notebook feature which was very useful to associates in addition to its data workflows.
Microsoft environment does not have the scalability of Nintex; it is perfect for small and medium-sized companies, especially in environments where Microsoft environment is almost entirely used. Although Microsoft offers options to connect to other applications, its platform lacks the development and robustness that Nintex provides. Nintex not only covers Microsoft environments but also Google and other important platforms.
Teradata is way ahead of its competitor because of its unique features of ensuring data privacy and data never gets corrupted even in worst case scenario. In most cases, the data corruption is a major issue if left unused and it leads to important data being wiped off which in ideal case should be stored for 3 years
The scalability is really bottlenecked by the imagination of the user. I was able to make processes for my own personal usage, making my daily tasks easier. I was also able to make processes that affected hundreds of employees, making large standardization and efficiency gains. So either way, the system is used the same way, and I was the limiting factor.
It had a positive ROI in terms of reducing the amount of time spent on data movement and curation by associates
It had a positive ROI in terms of increasing the number of insights from structured data
It reduced the number of data entry and analysis errors by associates which led to a positive ROI in terms of efficiency and reducing time wasted by tracking down errors in data
People have woken up to the amount of overlap after mapping their processes.
People can be resistant to process changes. You need to have the support from above or support from the 'business' that you are process changing to be able to see the positive impacts.
Numbers talk. if you can get a general salary figure from your HR dept to show savings for 'employee bands', then when you present reports, they will be all the richer in data.
Moving to Teradata in the Cloud-enabled a level of agility that previously didn't exist in the organization. It also enabled a level of analytic competency that was not achievable using other options on the aggressive timeline that was required. We didn't want to settle for reinventing a wheel when we had a super tuned performance capable beast readily available in Teradata. Teradata lets us focus on our business rather than spending money and effort trying to design software or database foundations features on an open source or lower performance platform.