This article explains how to monitor and understand your experiments’ performance.
Instructions
View your experiment’s performance
- In your Google Ads account, click the Campaigns icon .
- Click the Campaigns drop down in the section menu.
- Click Experiments.
- Find and click the experiment that you want to check performance for.
- Review the “Experiment Summary table” and scorecard, then choose to Apply experiment or End experiment.
What the scorecard shows
- Performance comparison: This shows the dates for which your experiment's performance is being compared to the original campaign's performance. Only full days that fall between your experiment's start and end dates and the date range you've selected for the table below will show. If there is no overlap, the full days between your experiment's start and end dates will be used for "Performance comparison."
- By default, you’ll notice performance data for Clicks, CTR, Cost, Impressions, and All conversions, but you can select the performance metrics you want to view by clicking the down arrow next to the metric name. You’ll be able to choose from the following:
- Clicks
- CTR
- Cost
- Avg. CPC
- Impr
- All conv. (available only if you’ve set up conversion tracking)
- Conv. rate (available only if you’ve set up conversion tracking)
- Conversions (available only if you’ve set up conversion tracking)
- Cost / conv. (available only if you’ve set up conversion tracking)
- View-through conv. (available only if you’ve set up conversion tracking)
- Incremental queries (available for broad match experiments)
- The first line below each metric name shows your experiment’s data for that metric. For example, if you notice 4K below “Clicks,” that means your experiment’s ads have received 4,000 clicks since it began running.
- The second line shows an estimated performance difference between the experiment and the campaign.
- The first value shows the performance difference your experiment saw for that metric when compared to the original campaign. For example, if you notice +10% for Clicks, it’s estimated that your experiment received 10% more clicks than the original campaign. If there’s not enough data available yet for the original campaign and/or the experiment, you’ll notice “‑‑”.
- The second value shows that if you chose a 95% confidence interval then, this is the possible range for the performance difference that might exist between the experiment and the original campaign. For example, if you notice [+8%, +12%], it means that there might be anywhere from a 8% to 12% increase in performance for the experiment when compared to the campaign. If there’s not enough data available yet for the original campaign and/or the experiment, you’ll notice “‑‑”. You can pick your own confidence intervals (80% is the default confidence interval) and be able to understand your experiment metrics better with dynamic confidence reporting.
- If your result is statistically significant, you’ll also find a blue asterisk.
Tip
Point your cursor over this second line for a more detailed explanation of what you’re reviewing. You'll be able to view the following information:
- Statistically significant: This means your p value is less than or equal to 5%. In other words, your data is likely not due to chance, and your experiment is more likely to continue performing with similar results if it’s converted to a campaign.
- Not statistically significant: This means your p value is greater than or equal to 5%. These are some possible reasons why your data could be considered not statistically significant.
- Your experiment hasn’t had enough time to run.
- Your campaign doesn’t receive enough traffic.
- Your traffic split was too small and your experiment isn’t receiving enough traffic.
- The changes you’ve made haven’t resulted in a statistically significant performance difference.
- Whether or not your data was shown to be statistically significant, you’ll notice an explanation like the following to show the level of likelihood that the performance data was due to random chance: “There's a 0.2% (p-value) chance of getting this performance (or a larger performance difference) due to randomness. The smaller the p-value, the more significant the result.”
- Confidence interval: You’ll also find more details about the confidence interval for the performance difference with an explanation like the following: “There's a 95% chance that your experiment notice a +10% to +20% difference for this metric when compared to the original campaign.”
-
Finally, you’ll find the actual data for that metric for the experiment and the original campaign.
What you can do in the scorecard
- In the scorecard, you can change the metric you find using the drop-down next to the metric.
- To review the scorecard for ad group in the experiment, click an ad group from the table below.
- To view details like the name and budget of a campaign, hover over the corresponding cell in the table.
Understand the time series chart
The time series chart displays the performance of up to 2 metrics in your experiment and shows how they've changed over time in both treated and control campaigns. With this chart, you can compare the effects your experiments have on a particular metric and learn more about how it performs over time.
Incremental queries
Incremental queries are search terms that matched a broad match keyword in your experiment but weren’t matched by the account during the time period of the experiment. These queries have at least one conversion attributed to their click. This is applicable to your broad match experiments. You can view up to 5 top incremental queries, if available, in your Experiments home page and Report page.
What happens when you pause or remove Performance Max experiment campaigns
You can pause or end your control or treatment campaign at any time. To restart a paused experiment, simply use the "Resume" button to resume the campaigns or manually reactivate the campaigns.
- Pausing campaigns:
- If you pause either the control or treatment campaign, the experiment will be paused. 100% of the traffic will go to the remaining campaign that's active.
- If you pause both the control and treatment campaign, the experiment will be paused.
- Experiment status on end date:
- If the experiment reaches the end date, the Performance Max syncer will change Performance Max Experiment status to “ended” or “launched” regardless of auto-apply. However, if the experiment result is good and auto-apply is turned on, the auto-apply pipeline will auto graduate the “ended” experiments afterwards.
- If the experiment reaches the end date and the auto-apply is turned off, you’ll need to apply the changes manually.
- Removing campaigns:
- If you remove the control or treatment campaign, 100% of the traffic will go to the remaining campaign that's active.
- If you only remove the control campaign, the experiment will be launched.
- If you only remove the treatment campaign, the experiment will end.
- If you remove both the control and treatment campaign, the experiment will end.
- If you remove the control or treatment campaign, 100% of the traffic will go to the remaining campaign that's active.
Campaign status | Traffic after user-action | Experiment status before end date | Experiment status on end date | |||||
User action | Control | Treatment | Control | Treatment | ||||
Pause either campaign | Paused | Active | 0% | 100% | Paused | If auto-apply is enabled and the treatment arm has favorable results, the experiment will be Launched. Otherwise, the experiment will be Ended | ||
Active | Paused | 100% | 0% | Paused | ||||
Pause both campaigns | Paused | Paused | 0% | 0% | Paused | |||
Remove either campaign | Removed | Active | 0% | 100% | Launched | |||
Active | Removed | 100% | 0% | Ended | ||||
Remove both campaigns | Removed | Removed | 0% | 0% | Ended |
Apply or end an experiment
To apply an experiment on a campaign or end an experiment for any reason, click the "Apply" or "End" button in the lower right corner of the "Experiment summary" card above the time series chart.
Review or edit a campaign comparable to an existing Performance Max campaign
During the experiment, you’ll need to manually edit comparable campaigns, otherwise Google will choose the comparable campaigns for you. After the experiment has ended, you’ll be unable to add or remove your comparable campaigns.
- View results alongside comparable campaigns: To determine how your experiment performs compared to similar campaigns, enable the "View Results with Comparable Campaigns" toggle in Google Ads reporting.
- Edit comparable campaigns (until experiment ends): You have complete control over which campaigns are considered "comparable" throughout the experiment. Simply edit your selections within your Performance Max experiments. If you don't manually choose comparable campaigns, Google will automatically select them for you. Once the experiment ends, you won't be able to add or remove comparable campaigns from the results.