Dynatrace is an APM scaled for enterprises with cloud, on-premise, and hybrid application and SaaS monitoring. Dynatrace uses AI-supported algorithms to provide continual APM self-learning and predictive alerts for proactive issue resolution.
$0
per synthetic request
ScienceLogic SL1
Score 9.1 out of 10
Enterprise companies (1,001+ employees)
ScienceLogic is a system and application monitoring and performance management platform. ScienceLogic collects and aggregates data across and IT ecosystems and contextualizes it for actionable insights with the SL1 product offering.
N/A
Pricing
Dynatrace
ScienceLogic SL1
Editions & Modules
Synthetic Monitoring
$0.001
per synthetic request
Kubernetes Platform Monitoring
$0.002
per hour for any size pod
Real User Monitoring
$0.00225
per session
Application Security
$0.018
per hour for 8 GIB host
Infrastructure Monitoring
$0.04
per hour for any size host
Full-Stack Monitoring
$0.08
per hour for 8 GIB host
No answers on this topic
Offerings
Pricing Offerings
Dynatrace
ScienceLogic SL1
Free Trial
No
No
Free/Freemium Version
No
No
Premium Consulting/Integration Services
No
Yes
Entry-level Setup Fee
No setup fee
Required
Additional Details
—
ScienceLogic SL1 offers four tiers:
SL1 Advanced – Application Health, Automated Troubleshooting and Remediation Workflows
SL1 Base – Infrastructure Monitoring, Topology & Event Correlation
SL1 Premium – AI/ML-driven Analytics, Low-Code Automated Workflow Authoring
SL1 Standard – Infrastructure Monitoring – with Agents, Business Services, Incident Automation, CMDB Synchronization, Behavioral Correlation
To get pricing for each tier, please contact the vendor.
Dynatrace provides us with details of every transaction - others only provided up us with snapshot samples of data.
ScienceLogic SL1
Verified User
Team Lead
Chose ScienceLogic SL1
As enterprise, we needed to leverage a tool that would resolve our problems for a fair price. Although ScienceLogic is not a fancy tool for some environments (as Dynatrace for the APM) or the visual appeal for observability (as Prometheus) we needed a tool that would broadly …
These are two different type's of products really that cross over each other at some intersections. Dynatrace is a Deep Diving monitoring tool as ScienceLogic monitors what can be seen on the surface, like logs, cpu memory, cpu and memory would be a crossover, and ScienceLogic …
From a monitoring perspective, in my opinion, ScienceLogic SL1 meets if not exceeds in all areas. The two caveats are APM and Network Flow Analytics (NFA) where ScienceLogic SL1 does not have a native capability but does integrate well with the leading APM and NFA providers.
ScienceLogic SL1 supports large scale of IT Infrastructure devices and vendors. Was the single tool providing multiple functionalities at same time and allowed to remove additional legacy tools used for monitoring. Allowed integration with incident management and CMDB. Allowed …
Verified User
Manager
Chose ScienceLogic SL1
From a capability perspective they stack up very similar but from a look and feel, ScienceLogic SL1 one is miles behind the curve on all three. We chose SL because we already had elements of the service in place on our infrastructure from our previous MSP so they were a …
Agentless, easy, and powerful. Easy to deploy, maintain and manage. The one-stop solution replaces multiple monitoring tools, has an inbuilt ticketing system, and can be integrated with any tools. Easy to deploy any application monitoring via REST API calls. Dynamic …
ScienceLogic SL1 has great UI to configure any monitoring after setup, also has agent-less monitoring that's the great feature against mandate agent monitoring tool like ITM and Zabbix. ScienceLogic SL1 UI has great flexibility to configure any monitoring and modification as …
Agentless monitoring is the best part of ScienceLogic SL1 monitoring, which is asked for by all the customers. UI is good and easy to handle. Port status from ScienceLogic SL1 collectors is easy even with bulk servers. The overall product is good for basic INFRA monitoring. …
Dynatrace is well suited to a number of tasks. It is important to determine who the end users are and gather good information to tailor their experience accordingly. For instance, business/marketing should not have access to some of the more technical data, and business metrics can be a distraction for IT operations personnel.
Appropriate if you are setting up a monitoring suite in new Infrastructure Environment. Definitely NOT suited for Migration Projects. ScienceLogic SL1 cannot cater to a lot of monitoring requirements which already would have been configured in old monitoring suite. Plus, limited support for customizations and having to go to "Feature Requests" route makes in extremely complicated.
We loved Dynatrace's ability to show the data flow - from the front end points through the back end points straight to the database and various API's. It was advanced in its data visualization. This is useful for debugging - showing when/where the errors are. It can even enable non-technical individuals in the corporation to help debug
Dynatrace has some great highly customizable integration options as well as monitoring. You can configure your layout & integration options to create custom monitoring alerts for your applications performance. Further you can increase the extensibility of using a REST API on your architecture.
Some advanced dev-ops systems are utilizing Kubernetes/docker aswell as Node.JS - Dynatrace was able to log and help understand all of our dev-ops needs. It gave us native alerts based off of deviations from the baseline that we set during initial configuration. These metrics are priceless.
Dynatrace does not monitor easily on a C-based application.
The way DPGR is addressed by Dynatrace is not very complete, and not clear. One thing is to mask the IP and request attributes but is not enough, the replay session feature is great but raises serious questions about user tracking.
Dashboards are quite old and are of Iron age. Need to have AP2 dashboards only instead of AP1 and consistent new design across all functionalities.
Reporting is not improved since Y2020 and need to revamp completely. Need to integrate Dashboards and Reporting. PowerBI Like functionality to be given OOTB. Reports should be extracted in Excel, PDF, HTML and should be heavily automated.
Create and Open APIs for basic and advanced monitoring data extraction.
Topology based Event Correlation and Suppression should be improved drastically. Need to identify critical network interfaces based on Topology and monitor them. Basic customization of Dynamic App and/or Powerpack to exclude/include certain metrics/events to be permitted OOTB instead of customizations.
Integration with ServiceNow to be improved and to be taken to next level. Automation Powerpack should be made available OOTB as part of base product and to be priced attractively.
Take product to next level where we can monitor actual impacted IT or Business Service instead of metrics and events BSM and Topology map to be auto discovered and identify the network dependencies and alternate paths automatically instead of manual creation of BSM.
We have already renewed our purchase with the company. They make it easy for us to get a temporary license for our contingency site that is only used for testing twice a year. We are expanding our license with for this tool. We find it very useful and will renew it again.
It is simply because of all the best possible autonomy solutions it is providing and getting better day by day. Using AI and Devops along with handy automation, The monitoring and Management of devices becomes much easier and the way it is growing in all the aspects is one the best reasons too. Evolution of the SL1 platform in the autonomy monitoring and management is quite appreciable.
Dynatrace is great to use once you understand how to use it correctly and get used to the layout of it. While I do not actively use it every day, whenever I do use it, I do have to get refamiliarized with it. However, once you have your dashboards setup correctly with the data that you want to see when you first login to Dynatrace, it's amazing.
The core functions are there. The complexity is due to the complexity of the space. The score is based on comfort (I no longer notice the legacy UI) and the promise that I see in the 8.12 Unified UI (a vast improvement). It is also based on the fact that with 8.12, you can now do everything in the new UI but you still have the legacy UI as a fallback (which should now be unnecessary for new installations)
SL is always there and online when you need to get info from it. The only times when SL was not available in our own data center, was when network links from out side of the data center was down and those links were not in our controll. Having a central database and people accessing it all over the world, may put a bit of constarin on the performance of the dashboards when reports gets generated, but that is far and few n between.
SceinceLogic SL1 architecture helps the platform to give a top-notch performance in every respect, Data collection to reporting happens very smoothly. With the new user interface pages load much faster. Individual appliances carrying the individual task ensure things are working without lag. Integration with ticketing tool(SNOW) is well managed by the ScienceLogic, no issue or much delay has been observed while interacting with an external tool.
Given that Dynatrace has become an informal industry standard, the plethora of information available on forums is massive. Most problems or roadblocks you come across are most likely (almost certainly, in fact) already solved and solutions available on these forums. The tech support at Dynatrace is also quite good, with prompt and knowledgeable people at their end.
So far, it's good as part of my overall experience, except for a couple of use cases. The support team is well knowledgeable, has technical sound, and is efficient. When support escalates to engineering, the issue gets stuck and takes months to resolve.
It was good, Do the online training first and understand it and you will get the most out of the in-person training that way. This also takes you to an advanced level which is very good and the training as been overhauled once again along with new product coming in such as Zebruim / Skylar, worth going through again if it a while back that you first did this.
There are a lot of educational materials and courses on the SL1 training site (Litmos university). However the recording quality is sometimes not very good - screen resolution is low. There is a lack of professional rather than user-oriented documents and there are mistakes in documentation and education is not well structured.
As first time developers, getting to grips with powerpack development using SNMP, Powershell and Python etc, was not helped by poor and badly organised online documentation. In many cases, we had to look at existing powerpacks and try to work out what it was doing and why - not always with much success. Even after receiving expert level training, the development of some powerpacks would not have been possible without access to the SL1 support staff.
Synthetic Monitoring automatically does what other products do only through the use of other tools or through the development of user applications that still have a high cost of maintenance. The other products are not immediately usable and require many customizations. Through the use of configuration automatisms, you can be immediately operational and, in our case, we detected several imperfections in the applications.
Science logic SL1 is so user friendly and it's really easy to navigate between function. I would recommend Sciene logic SL1 to all of them who are looking for really useful monitoring tool and expecting easy way of managing it.
Our deployment model is vastly different from product expectations. Our global / internal monitoring foot print is 8 production stacks in dual data centers with 50% collection capacity allocated to each data center with minimal numbers of collection groups. General Collection is our default collection group. Special Collection is for monitoring our ASA and other hardware that cannot be polled by a large number of IP addresses, so this collection group is usually 2 collectors). Because most of our stacks are in different physical data centers, we cannot use the provided HA solution. We have to use the DR solution (DRBD + CNAMEs). We routinely test power in our data centers (yearly). Because we have to use DR, we have a hand-touch to flip nodes and change the DNS CNAME half of the times when there is an outage (by design). When the outage is planned, we do this ahead of the outage so that we don't care that the Secondary has dropped away from the Primary. Hopefully, we'll be able to find a way to meet our constraints and improve our resiliency and reduce our hand-touch in future releases. For now, this works for us and our complexity. (I hear that the HA option is sweet. I just can't consume that.)