Dev guideAPI Reference
Dev guideAPI ReferenceUser GuideGitHubNuGetDev CommunitySubmit a ticketLog In
GitHubNuGetDev CommunitySubmit a ticket

Updated Optimizely application UI

Explains the differences in the Optimizely application user interface between Full Stack Experimentation legacy and Feature Experimentation.



For a full side-by-side comparison of the Full Stack Experimentation (legacy) and Feature Experimentation refer to our user documentation page on migration examples.

Experiments and Features into one tab

The Experiments and Features tabs were replaced with one Flags tab. This is the central dashboard for your flags and their rules which support targeted deliveries and experiments within an environment in your project. Everything starts with a flag, whether running an experiment or rolling out a feature.

Full Stack Experimentation - Legacy

Experiments tab

Experiments page

Features tab

Features page

Feature Experimentation

New flags overview page

New flags overview page

Flag details

The new Flag Details page is where you can manage a flag's ruleset. The ruleset uses sequential logic for delivering your feature or experiments. Aside from all experiments now being tied to a flag, all other components for creating experiments (audiences, metrics, and variations) remain the same as in the Full Stack (Legacy) experience.

You can create multiple rules for the same flag, providing flexibility in the experiences you deliver to users. Rules are evaluated sequentially, allowing you to maximize customer experience, given your priorities.

For example, you can prioritize your qualifying traffic to a flag for an experiment and delegate non-qualifying traffic to a default experience or targeted delivery. Because all experiments are based on flags, all winning variations in Feature Experimentation can be rolled out without needing to go through your entire software development lifecycle, including an extra deployment.

New flag detail page

New flag details page

Shared variations

Variations and Variables are now shared across all rules in a flag, making it easy to reuse variations across your rules.

new shared variations page

New Variations page

Different results per environment

In Full Stack (Legacy), a single experiment exists across all environments and shares one results page. When you run quality assurance (QA) on an experiment, those events get mixed with the live production results. In Feature Experimentation, experiment rules are scoped by the environment, so results are also scoped by the environment. This eliminates a common pain point wherein events created in QA are mixed with live production results and the need to reset results before running the experiment in your primary environment. Refer to Migrate existing experiment results for information.

Reports page

In Full Stack Experimentation (Legacy), experiment results can only be viewed at the experiment level, requiring clicking view results for each experiment and not providing a centralized view of your results.

Feature Experimentation introduces a new Reports tab, where you can access results for your experiments from one view. Reports are displayed for one environment at a time and can be changed with the filter options.

new reports page

New Reports page