Google Ad Manager

How to run a manual experiment in Google Ad Manager

The one-size-fits-all approach to advertising never cuts it. 

Your ability to create user-friendly ads depends on how well you experiment with different formats, size, color combinations, and placements. 

Your ads may be good enough to prevent policy violations in Google ad Manager, but if you want incremental growth, there’s always room for customization and experimentation. 

Continue reading to find out how you can use Google’s new manual experiment feature to capture more revenue.

What is a Manual Experiment in Ad Manager?

Google has developed a new tool called ‘Manual Experiments’ to help publishers run their own tests or experiments in Google Ad Manager. 

With this new feature, publishers will be able to create and run experiments and analyze results within the existing GAM framework.

In other words, the manual experiment feature helps publishers maximize their revenue with manual opportunities for improving yield.

Google defines manual experiments as:

“An experiment you define based on your own criteria and schedule. The experiment runs on a percentage of your network’s actual traffic to test how applying those settings would impact revenue.”

Google

Publishers can run up to 100 active experiments within their existing ad manager network at any given time. Active experiments include trails that are running, completed, paused, or waiting for you to take action.

Learn more:

1. Google Ad Manager 360: Everything You Get with GAM 360

2. How to Set Up Multi-format Ads in GAM (Google Ad Manager)

Manual Experiments Type: 

As of now, Publishers can run three types of experiment on the ad manager: 

Manual experiments in Ad Manager

1. Category block experiments:

Publishers can run experiments to unblock categories from their existing protections and increase coverage. 

Through category block experiments, publishers can analyze the impact of running specific ad types on currently blocked sites. 

2. Native ad style experiments: 

Publishers can experiment with two different visual elements and native ad styles to see which one will perform better. 

Native ad style experiments can only compare native styles on the existing native placements. You cannot compare multi-format ads in the same placement. 

3. Unified pricing rule experiments: 

This manual experiment type allows publishers to test changes to unified pricing rules, which helps them understand their impact on the publisher’s exchange inventory. 

In other words, publishers can review the impact of increasing or decreasing price floor on specific inventory. 

Note: Google plans to roll out more experiment types in the future based on user feedback and demand. 

Learn More:

1. How to Run Video Ads in Google Ad Manager (GAM)

2. Guide to Creating and Running Native Ads in Google Ad Manager

How does it work?

Manual experiment feature is built on the Ad manager’s existing Opportunities and Experiments tool, which uses Google’s experiment framework and machine learning to automatically add suggestions for publishers.

In the ad manager, you can check the list of available manual experiment types and review your active experiments by clicking on Home > Reporting > Experiments.

You can conduct manual experiments by selecting the experiment type and criteria of your choice. The tool asks you to define an experiment trial before running it for a specified period. 

You can compare impression traffic allocated to the ‘control’ and ‘experiment’ groups to see which performed better during the experiment. 

You can also run an experiment from an opportunity suggested by the Ad Manager. 

Furthermore, the tool allows you to run additional trails if needed and apply experiment settings to your ad manager network. 

How to run a manual experiment in GAM:

Follow these seven steps to run a manual experiment in Google Ad Manager:

  1. Sign in to your GAM account. 
  2. On the dashboard, click reporting
  3. Click Experiments > New experiments from the experiment type you prefer 
  4. Enter the name of your experiment 
  5. Enter a start and end date for the experiment trial 
  6. Set the percent of impression traffic that you want to allocate to experiment 
  7. Click save 

Note: Each experiment needs to run for a minimum of seven days to get conclusive results. Publishers can set a trial to start immediately after saving the settings or schedule it for later. 

The maximum duration for the experiment is 31 days. The trial will expire after this period. Once the trial ends, you can evaluate the results to take the following steps:

  • Apply experiment result as an opportunity 
  • End the experiment and keep original settings 
  • Run another trail to find better opportunities

Learn More:

1. Key Features of the New Google Ad Manager Interface

2. How to Generate Reports in Google Ad Manager

Manual experiment best practices:

Google recommends the following practices for publishers who want to get the most out of manual experiments and capture their potential revenue on market behavior: 

Best practices for manual experiments in ad manager

1. Run experiments for a longer duration

It’s important to run these experiments over a long period to understand audience behavior and study the impact of changes you’ve made. 

As mentioned above, the maximum duration for running a trial is 31 days, and the minimum is 7. Ideally, publishers need 2-3 weeks to get a grasp of behavior changes. 

2. Comparison between different settings:

You need to make a thorough comparison of both settings to get an understanding of audience behavior. 

Make sure to compare impression traffic allocated to the ‘control’ and ‘experiment’ groups to see which performed better during the experiment. 

3. Consider total revenue: 

In addition to comparing behavior changes in the experiment and control group, you should look at the numbers too because experiment settings are likely to impact your total revenue. 

You should verify whether the expected revenue in the experiment group is in line with your initial expectations against your network as a whole. 

4. Experiment more often 

Experiments are very likely to encourage behavior change amongst your target audience. Google recommends publishers run more trials to get a deeper understanding of audience behavior. 

It’s better to experiment with both lower and higher traffic allocation percentages. While the lower traffic allocation might not reflect big behavioral changes, they’re also less risky (if the trial has a negative impact). 

We suggest allocating traffic percentages from lower to higher. Once you find your trials performing at an acceptable level, you can start another test with the same settings at a higher traffic allocation to measure the effects. 

Learn More:

1. How to Use Google Ad Manager with AdSense and Ad Exchange

2. How does Open Bidding Work in Google Ad Manager?

Publishers have started experimenting:

Publishers like the Atlantic, UOL, and Storer have already started using manual experiments in Google Ad Manager, and they see success. 

In fact, publishers who used manual experiments during the beta period achieved a 6.5% revenue increase. 

Similarly, in 2020, publishers who used Opportunites and Experiments tools accepted over 2000 opportunities and gained 3.5% more revenue. 

“We’ve been able to seamlessly run experiments using Manual Experiments – giving us insight into the impact of different floor prices and helping us determine how to effectively price floors to earn the most revenue. We’re excited to see how this feature evolves.”

– Ryan McConaghy, The Atlantic. 

The new addition to GAM is helping publishers test pricing rules as well: 

“With Manual Experiments, we now have an important tool to speed up and simplify the testing and optimization process in our unified price rules decisions.”

Fabricio Gomes, UOL.
Publishers Secrets for Higher Ad Revenue with IZooto

In Closing: 

We’ve picked on manual experiments for a reason – in today’s dynamic environment, the best way to improve revenues is to test and experiment. 

Ads often become a victim of banner blindness and poor UX. It’s time to scrap your traditional ads.

We encourage you to keep A/B testing individual elements of your ads to see what clicks with your audiences. 

Ready to start manual experiments? Get a demo of AdPushup to find out how to use its built-in ad optimization and experimentation features for more ad revenue.


Write A Comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.