Optimizely Feature Experimentation unites feature flagging, A/B testing, and built-in collaboration—so marketers can release, experiment, and optimize with confidence in one platform.
N/A
Sentry
Score 8.8 out of 10
N/A
Sentry provides engineering teams with tools to detect and solve user-impacting bugs and other issues.
Based on my experience with Optimizely Feature Experimentation, I can highlight several scenarios where it excels and a few where it may be less suitable. Well-suited scenarios: - Multi-Channel product launches - Complex A/B testing and feature flag management - Gradual rollout and risk mitigation Less suited scenarios: - Simple A/B tests (their Web Experimentation product is probably better for that) - Non-technical team usage -
Great for standard web application performance monitoring, analytics and error reporting. Shows line level code errors, gives insight into performance issues (plugins, API issues, etc.). Automation and scheduled scanning in production gives client visibility into 'after deployment' value. Also lets a relatively small number of developers keep tabs on a handful of different site/applications without needing a bunch of tools. The UI is pretty complicated and can be overwhelming for new users. Documentation could be better for the learning curve,
It is easy to use any of our product owners, marketers, developers can set up experiments and roll them out with some developer support. So the key thing there is this front end UI easy to use and maybe this will come later, but the new features such as Opal and the analytics or database centric engine is something we're interested in as well.
Great web interface. Lots of data available in a really clean format, with filtering options and more.
Per-user exception tracking. User is complaining about something being broken? Look up their account ID in Sentry and you can see if they've run into any exceptions (with device information included, of course).
Source map uploading. Took a little while to figure this out but now we have our deploy script upload sourcemaps to Sentry on each deployment, meaning we get to see stack traces that aren't obfuscated!
Very generous free tier – 10,000 events per month. We're nowhere near that yet.
Would be nice to able to switch variants between say an MVT to a 50:50 if one of the variants is not performing very well quickly and effectively so can still use the standardised report
Interface can feel very bare bones/not very many graphs or visuals, which other providers have to make it a bit more engaging
Doesn't show easily what each variant that is live looks like, so can be hard to remember what is actually being shown in each test
Easy to navigate the UI. Once you know how to use it, it is very easy to run experiments. And when the experiment is setup, the SDK code variables are generated and available for developers to use immediately so they can quickly build the experiment code
Its incredibly versatile, but that leads to complexity for the uninitiated, which can be intimidating. Nevertheless its a well polished product, in our case leading to only using it for a focus on frontend is still more cost effective than buying a one-to-rule-them-all tool...
When Google Optimize goes off we searched for a tool where you can be sure to get a good GA4 implementation and easy to use for IT team and product team. Optimizely Feature Experimentation seems to have a good balance between pricing and capabilities. If you are searching for an experimentation tool and personalization all in one... then maybe these comparison change and Optimizely turns to expensive. In the same way... if you want a server side solution. For us, it will be a challenge in the following years
It is cheaper and offers better support for front-end applications for enterprise large environments with more then 30 scrum teams and hundreds of micro frontend applications. The configuration options, both with the agent and from the user interface, are superior to other tools, and the documentation is also very easy to use.
We have a huge, noteworthy ROI case study of how we did a SaaS onboarding revamp early this year. Our A/B test on a guided setup flow improved activation rates by 20 percent, which translated to over $1.2m in retained ARR.