A/B testing campaigns can provide valuable insight into your audience. Klaviyo makes determining what changes you should make to your sending strategy simple. Not only will you know the winning variation for an A/B test, but also other important factors, like the win confidence percentage and winning metric details. In this article, we’ll go over what you can find on the A/B test results page as well as what each of these indicators mean.
A/B Test Results
To see the results of an A/B test, navigate to the campaign you are testing. Next, select the A/B Test Results tab of the main campaign report.
This page is broken into two parts:
- Test Overview
The sections below break each of these down.
In the Test Overview section, you’ll see a snapshot of the test results, including the:
- Winning variation, win confidence, and statistical significance
- Winning metric
- Test size
Winning Variation, Win Confidence, and Statistical Significance
On the left-hand side, you’ll see the winning variation, if the test is finished, or the variation currently in the lead for any ongoing A/B test.
Below this, you’ll see the win confidence. This is how likely it is that the winning variation truly had the highest campaign open or click rate after taking random chance into account.
Say, for instance, you want to see if adding emojis to your subject line impacts your open rate with your newsletter list. If you only send the test to two subscribers and the emojis have a better open rate, you wouldn’t say conclusively that the rest of your newsletter list would feel the same way. If instead you tested with 100 people in your newsletter list and found that emojis have a substantially higher open rate, you would feel much more confident that using emojis in your subject lines is better for your audience.
Under the win confidence, you’ll see whether or not the variation shown is statistically significant. If it is deemed significant, this means that the variation has a high win probability. For Klaviyo campaigns, this means the variation has a 90% chance or more of winning. In this case, a green tag saying Statistically significant will appear. On the other hand, if the variation is not statistically significant, you will see a gray tag saying Not statistically significant. Nothing will appear if the results are inconclusive. For details on how the statistical significance of a campaign is decided, read this article on how statistical significance is determined.
The Test Overview section will also show the status for the test; i.e., whether it is still in progress or complete. If the test is complete, the date and time when it finished is also displayed.
The winning metric, which is the metric you choose when you start an A/B test, is shown here as well. This area will also show the percent difference between the leading and next-best variation.
With the release of iOS15, macOS Monterey, iPadOS 15, and WatchOS 8, Apple Mail Privacy Protection (MPP) changed the way that we receive open rate data on your emails by prefetching our tracking pixel. With this change, it’s important to understand that open rates will be inflated.
With regard to A/B testing, we anticipate that our tools should account for these inflated open rates; however, you may need a higher threshold to reach statistical significance. If you regularly conduct A/B testing AND have greater than ~45% of opens on Apple Mail, we suggest creating a custom report that includes an MPP property. You can also identify these opens in your individual subscriber segments.
For complete information on MPP opens, visit our iOS 15: How to Prepare for Apple’s Changes guide.
As the name suggests, the test size is the size of the test group for a certain variation. You’ll see this as both a percentage and the actual number of recipients.
In the Variations card, you’ll see the following (from left to right):
- The variation subject line, preview text, and sender name and email address
- Number of recipients
- Open rate and number of opens
- Click rate and number of clicks
- Chosen conversion metric and number of recipients who converted via that metric
For example, the default conversion metric is Placed Order. If you use this, the card will show the placed order rate, the number of recipients who placed an order, and the average order value
12 hours after the campaign sends — or, if you chose to run an A/B test with a test period, 12 hours after the test period ends — you’ll see a trophy icon next to the subject line of the winning variation.
Check out this strategy guide: Best Practices for A/B Testing
Learn about other A/B tests you can run: