Optimizely Feature Experimentation vs. Split

Overview
ProductRatingMost Used ByProduct SummaryStarting Price
Optimizely Feature Experimentation
Score 7.5 out of 10
N/A
Optimizely Feature Experimentation combines experimentation, feature flagging and built for purpose collaboration features into one platform.N/A
Split
Score 5.0 out of 10
N/A
Split provides a feature flagging platform that separates code releases from feature deployment, allowing DevOps teams to test new features while still in production. Split offers compatibility with a wide range of programming languages and third party software integrations such as Slack and Jira Software.N/A
Pricing
Optimizely Feature ExperimentationSplit
Editions & Modules
No answers on this topic
No answers on this topic
Offerings
Pricing Offerings
Optimizely Feature ExperimentationSplit
Free Trial
NoYes
Free/Freemium Version
YesYes
Premium Consulting/Integration Services
YesNo
Entry-level Setup FeeRequiredNo setup fee
Additional Details
More Pricing Information
Community Pulse
Optimizely Feature ExperimentationSplit
Ask people about this product

See helpful people who have experience with this product

Top Pros

No answers on this topic

Top Cons

No answers on this topic

Best Alternatives
Optimizely Feature ExperimentationSplit
Small Businesses
Kameleoon
Kameleoon
Score 9.5 out of 10
Kameleoon
Kameleoon
Score 9.5 out of 10
Medium-sized Companies
Kameleoon
Kameleoon
Score 9.5 out of 10
Kameleoon
Kameleoon
Score 9.5 out of 10
Enterprises
Kameleoon
Kameleoon
Score 9.5 out of 10
Kameleoon
Kameleoon
Score 9.5 out of 10
All AlternativesView all alternativesView all alternatives
User Ratings
Optimizely Feature ExperimentationSplit
Likelihood to Recommend
7.4
(20 ratings)
-
(0 ratings)
Likelihood to Renew
8.0
(1 ratings)
-
(0 ratings)
Usability
9.0
(1 ratings)
-
(0 ratings)
Implementation Rating
10.0
(1 ratings)
-
(0 ratings)
Product Scalability
5.0
(1 ratings)
-
(0 ratings)
User Testimonials
Optimizely Feature ExperimentationSplit
Likelihood to Recommend
Optimizely
Optimizely Feature Experimentation works really well for setting up feature flags with an easy UI for turning them on and off or ramping up a gradual rollout. It also works really well to set up split tests where you can split your traffic by percentage as well as almost any custom data attribute you wish to define. This is more for robust features and less for visual changes - Optimzely Edge or Web are better suited for that.
Read full review
Split
No answers on this topic
Pros
Optimizely
  • Its ability to run A/B tests and multivariate experiments simultaneously allows us to identify the best-performing options quickly.
  • Optimizely blends into our analytics tools, giving us immediate feedback on how our experiments are performing. This tool helps us avoid interruptions. With this pairing, we can arrive at informed decisions quickly.
  • Additionally, feature toggles enable us to introduce new features or modifications to specific user groups, guaranteeing a smooth and controlled user experience. This tool helps us avoid interruptions.
Read full review
Split
No answers on this topic
Cons
Optimizely
  • Splitting feature flags from actual experiments is slightly clunky and can be done either as part of the same page or better still you can create a flag on the spot while starting an experiment and not always needing to start with a flag.
  • Recommending metrics to track based on description using AI
Read full review
Split
No answers on this topic
Usability
Optimizely
All features that we used were pretty clear. They have a good documentation
Read full review
Split
No answers on this topic
Implementation Rating
Optimizely
It’s straightforward. Docs are well written and I believe there must be a support. But we haven’t used it
Read full review
Split
No answers on this topic
Alternatives Considered
Optimizely
Optimizely Feature Experimentation is better for building more complex experiments than Optimizely Web. However, Optimizely Web is much easier to kickstart your experimentation program with as the learning curve is much lower, and dedicated developer resources are not always necessary (marketers can build experiments quickly with Optimizely Web without developers' help).
Read full review
Split
No answers on this topic
Scalability
Optimizely
had troubles with performance for SSR and the React SDK
Read full review
Split
No answers on this topic
Return on Investment
Optimizely
  • Experimentation is key to figuring out the impact of changes made on-site.
  • Experimentation is very helpful with pricing tests and other backend tests.
  • Before running an experiment, many factors need to be evaluated, such as conflicting experiments, audience, user profile service, etc. This requires a considerable amount of time.
Read full review
Split
No answers on this topic
ScreenShots

Optimizely Feature Experimentation Screenshots

Screenshot of AI Variable suggestions: AI helps to develop higher quality experiments. Optimizely’s Opal suggests content variations in experiments, and helps to increase test velocity  and improve experiment qualityScreenshot of Integrations: display of the available integrations in-app.Screenshot of Reporting used to share insights, quantify experimentation program performance using KPIs like velocity and conclusive rate across experimentation projects, and to drill down into the charts and figures to see an aggregate list of experiments. Results can be exported into a CSV or Excel file, and KPIs can be segmented using project filters, experiment type filters, and date rangesScreenshot of Collaboration: Centralizes tracking tasks in the design, build, and launch of an experiment to ensure experiments are launched on time . Includes calendar, timeline, and board views in customizable views that can be saved to share with other stakeholdersScreenshot of Scheduling: Users can schedule a Flag or Rule to toggle on/off,  traffic allocation percentages, and achieve faster experimentation velocity and smoother progressive rolloutsScreenshot of Metrics filtering: Dynamic event properties to filter through events. Dynamic events provide better insights for experimenters who can explore metrics in depth for more impactful decisions

Split Screenshots

Screenshot of the Targeting Rules interface, used to progressively deliver new features to market. First roll out to internal users, beta testers, then customer segments.Screenshot of the interface to run A/B tests for any feature released using customizable metrics from any data source.