The Reports page helps you interpret experiment metrics using Optimizely's Stats Engine, a unique approach to statistics for the digital age developed in conjunction with leading researchers at Stanford University. Stats Engine embeds innovations combining sequential testing and false discovery rate control to deliver speed and accuracy for businesses making decisions based on real-time data.
Once you run an experiment, dig into the Reports page to learn how users respond to your experiment.
For more information, see these Optimizely Knowledge Base articles:
- For getting to meaningful stats sooner, use the Stats Accelerator.
- How Optimizely calculates results.
Optimizely A/B tests do not currently support retroactive results calculation.
When you use deliveries, Optimizely doesn't send an impression to the Results page. This is because deliveries are for rolling out flags that are already tested in experiments. To measure the impact of a feature flag, run an experiment instead.
Segment your results to see how different groups of users behave, compared to users overall.
By default, Optimizely shows results for all users who enter your experiment. However, not all users behave like your average users. Optimizely lets you filter your results so you can see if certain groups of users behave differently from your users overall. This is called segmentation.
For example, imagine you run an experiment with a pop-up promotional offer. This generates positive lift overall, but when you segment for users on mobile devices, it's a statistically significant loss. Maybe the pop-up is disruptive or difficult to close on a mobile device. When you implement the change or run a similar experiment in the future, you might exclude mobile users based on what you've learned from segmenting.
Segmenting results is one of the best ways to gain deeper insight beyond the average user's behavior. It's a powerful way to step up your experimentation program.
Experiments don’t include “out-of-the-box” attributes like browser, device, or location attributes (these attributes are included in our Web project). Optimizely’s SDKs are platform-agnostic: we don't assume which attributes are available in your application or what the format is. In experiments, all segmentation is based on the custom attributes that you create.
Here's how to set them up:
- Create the custom attributes that you want to use for Results page segmentation.
- Target audiences based on your custom attributes.
- Pass the custom attributes into the user context for your experiment or app. For an example, see the Create User Context topic in your language's SDK reference.
After you define custom attributes and pass them in the SDK, they'll be available as segmentation options from the drop-down menu at the top of the Results page.
You can segment your entire Reports page or the results for an individual metric. Segmenting results helps you get more out of your data by generating valuable insights about your user.
- Navigate to your Reports page.
- Click Segment and select an attribute from the dropdown.
Under the Segment dropdown, you'll find default segments and custom segments all in one place.
For Optimizely to use attributes for segmentation, the attribute must be defined in the datafile, and it must be included in the user context object. However, it does not have to be added as an audience to the test.
When segmenting results, a user who belongs to more than one segment will be counted in every segment they belong to. However, if a user has more than one value for a single segment, the user is only counted for the last-seen value they had in the session.
See our KB article on how visitors and conversions are counted in segments.
- Discrepancies in third-party data
- How long to run an experiment
- How Optimizely counts conversions
- Interpret your results
- Stats Accelerator: Use algorithms to boost results
- Take action based on the results of an experiment
- Tracking offline conversion events with Optimizely Classic and Optimizely X
- Integrate Optimizely data with your own app usage data and metrics by enriching events with event tags and by exporting data.
Updated about 1 year ago