A/B Testing Ads on Facebook & Instagram

Growth Intelligence
0 min read
June 5, 2024

A/B testing is a powerful method to improve ad performance on Facebook and Instagram. By comparing two versions of an ad strategy, marketers can determine which one resonates better with their audience. Using A/B tests, businesses can optimize variables such as ad images, text, audience, or placement to increase engagement and conversions. For instance, testing vertical videos versus horizontal videos can lead to significant cost savings per web conversion.

Setting up an A/B test in Meta's Ads Manager is straightforward and can yield insightful results. This tool allows businesses to use an existing campaign as a template, making the process efficient and user-friendly. Ad campaigns should run for at least seven days to ensure reliable results, with a maximum duration of 30 days best practices for A/B tests.

When running an A/B test, it’s crucial to test only one variable at a time. This approach helps in accurately determining what drives better performance. For example, small businesses have found success by testing creative variables first, like different video formats or static ads A/B Testing Ads.

Executing A/B Tests on Social Platforms

Performing A/B tests on social platforms like Facebook and Instagram allows marketers to pinpoint which ads resonate most with their audience. By comparing different versions of ads, advertisers can optimize their campaigns for better performance.

Setting Up A/B Tests for Facebook Ads

To execute an A/B test on Facebook Ads, start in the Ads Manager. Use the A/B Testing tool to test different variables like ad images, text, audience, or placement. This tool ensures that each ad version is shown to different audience segments, preventing overlap.

Steps to set up A/B tests:

  1. In Ads Manager, choose your campaign and click on "A/B test."
  2. Select the variable you want to test. It could be an image, headline, or target audience.
  3. Define the duration of the test. Facebook recommends a minimum of 7 days for accurate results.
  4. Analyze the performance. Metrics such as click-through rate (CTR) and conversion rates help determine the winning version.

For detailed guidelines, refer to Meta's resources on A/B Testing Ads on Facebook.

Optimizing Instagram Ads through A/B Testing

Instagram also supports A/B testing within the Meta Ads Manager. Identifying the most engaging ad elements is crucial for maximizing reach and conversions.

Steps to optimize Instagram ads:

  1. Access Meta Ads Manager and create an A/B test for your Instagram campaign.
  2. Choose a single element to test, like the ad's visual content or call-to-action.
  3. Set up the test to run for a period that provides sufficient data, typically at least 7 days.
  4. Review specific metrics such as engagement rates, likes, and shares to evaluate performance.

For more insights, visit the Meta Business Help Center page on A/B testing.

Keep tests simple by changing only one variable at a time, ensuring cleaner and more precise results. This method not only refines your ad strategies but also provides valuable insights into audience preferences.

Analyzing A/B Testing Results

Understanding how to interpret and act on A/B testing results is crucial for improving ad performance on Facebook and Instagram. This section will cover how to read data and metrics from your tests and what steps to take based on the insights gathered.

Interpreting Data and Metrics

To start, locate your A/B test results within the Ads Manager, where you will see various metrics. Look for metrics like CTR, conversion rate, and cost per result. The CTR shows how often people who see your ad click on it. A higher CTR suggests that the ad is engaging.

Conversion rate indicates how many of those clicks lead to a desired action, such as a purchase. Cost per result helps understand the financial efficiency of the ad. High costs might indicate an issue with the ad's effectiveness or targeting. Be sure to compare these metrics between different versions of your ads to see which performs better.

Actionable Insights and Next Steps

Once you have interpreted the data, identify which ad version performed the best. For instance, if Ad A had a higher CTR but a lower conversion rate compared to Ad B, you may need to tweak the call-to-action or landing page of Ad A.

Use these insights to inform future campaigns. If a specific image or headline resonates more with your audience, incorporate similar elements in your next ads. Document these findings to create a knowledge base for future reference.

Making educated adjustments based on these results can lead to more effective ad campaigns and better return on investment (ROI). Check out the details on viewing A/B test results to refine your strategy.

Nate Lorenzen
Founder
Jenner Kearns
Chief Delivery Officer
Jenner Kearns
Chief Delivery Officer
Jenner Kearns
Chief Delivery Officer
Kenneth Shen
Chief Executive Officer
Kenneth Shen
Chief Executive Officer
Kenneth Shen
Chief Executive Officer
Kenneth Shen
Chief Executive Officer
Jenner Kearns
Chief Delivery Officer
Kenneth Shen
Chief Executive Officer
Jenner Kearns
Chief Delivery Officer
Jenner Kearns
Chief Delivery Officer
Jenner Kearns
Chief Delivery Officer
Jenner Kearns
Chief Delivery Officer
Kenneth Shen
Chief Executive Officer
Jenner Kearns
Chief Delivery Officer
Kenneth Shen
Chief Executive Officer
Kenneth Shen
Chief Executive Officer
Isla Bruce
Head of Content
Isla Bruce
Head of Content
Isla Bruce
Head of Content
Jenner Kearns
Chief Delivery Officer
Isla Bruce
Head of Content
Kenneth Shen
Chief Executive Officer
Isla Bruce
Head of Content
Isla Bruce
Head of Content
Isla Bruce
Head of Content
Kenneth Shen
Chief Executive Officer
Isla Bruce
Head of Content

Read next

Marketing Side Quest: Advertising on X

X, formerly Twitter, is many things but one thing it is not is discussed by the DTC Twitter community.

This is partially X's fault. Unlike every other ad platform they do zero paid marketing to attract SMBs to their platform.

In 2024, I've spent a lot of time with top level sales teams and engineers at X understanding the intricacies of this platform.

If X won't market their marketing, then maybe I can help.

Format

This guide will be broken into three sections: Start with Signal, Supercharge with Strategy, Accelerate with Creative.

When tackling advertising on any digital platform, it is paramount to go through these sections in sequence. A great strategy can be ruined by bad signal. Great creative can be ruined by bad strategy or signal.

Don’t just pass go. You need to go through each step purposefully.

Start with Signal: Unlock the Power of X's Ads

Unlike Meta, X's prediction algorithm isn't strong. (More on this later.)

To succeed on X, improving signal fidelity is key.

Let's dive into how to optimize it!

Enable First Party Cookies

Head over to Events Manager > Settings and turn on First Party Cookies. This simple step ensures X can track and use user interactions effectively.

Implement Conversion API

For serious advertisers, X offers a server-to-server integration called Conversion API. It’s a game-changer for scaling, but note: it requires a developer to implement.

Why Conversion API?
  • Provides a direct line of data from your server to X.
  • Ensures high-fidelity signals, crucial for optimizing your campaigns.
  • While X doesn’t popularize it, it’s mandatory for large-scale success. #ProTip
Leverage Custom Conversions

With a limited signal pool on X, custom conversions can be as effective as standard events for optimization. My preference? Standard events. But custom conversions can be a flexible alternative.

Custom vs. Standard Conversions: Which to Choose? ‍
  • Standard events are generally more straightforward.
  • Custom conversions offer more tailored optimization.
  • Evaluate your specific needs and choose accordingly.

Supercharge with Strategy: Optimize Your X Ads

Objectives: Soft Focus for Better Results

DTC advertisers often lean towards the Sales objective, which is great! But don’t ignore others. For driving leads, "Website Traffic" outperforms "Sales" tied to a lead event. Keep a soft focus and test different objectives.

Autobid: Make It Work for You

X’s cost control settings aren’t as effective as Meta’s. The key is making Autobid hit your targets. X's algorithm isn’t as strong, so this is crucial for cost control. #ProTip

Things to Turn Off: Optimize Placements

Unlike Meta & Google, on X, it’s best to turn off certain placements. Turn off profiles, search results, and replies. Keep the home timeline on for the best performance.

Follower Lookalike Audiences (LAL): Hyper-targeting

While Meta focuses on broad targeting, X excels with specific audiences. Build a follower LAL with no more than 6 hyper-related profiles. Use optimized targeting, but also test with it turned off to see what works best.

Campaign & Ad Set Structure: Keep It Organized

Use CBO (Campaign Budget Optimization) and limit ad sets in any campaign. Organize campaigns by creative theme and audience. Each ad set should relate to the theme but vary in creative type (e.g., statics vs. video). Ensure creative in each ad set is similar in tone and messaging.

Accelerate with Creative: Maximize Your X Ads Impact

Now that Signal and Strategy are covered, it's time to Accelerate all this hard work with Creative. 🎨 For X, think 1990s infomercial: BIG, LOUD, BOLD. Subtlety doesn't work here. Check out these examples in the linked post.

Be Bold, Be Loud

Use BIG, LOUD, BOLD typeface on either statics or videos. Stand out in the feed with eye-catching creative. #CreativeTips

Manage Toxicity

Communities on X can be toxic. Read your comments and consider turning off replies if they are highly negative. Protect your brand!

Best Practices Checklist
  • You’ve got 6 seconds: Persuasion on X happens fast. Most message recall is between 3-5 seconds. Use 1:1 or 16:9 to dominate the feed.
  • Reading on X doesn't have to be boring: Use artful captions and visual cues. These can make posts 11% more likely to be viewed and viewed 28% longer.
  • Brand your assets from 00:00: In eye-tracking data, the left-hand side of posts had 40% higher visibility and 30% quicker time to focus.
  • Keep it simple: A concise message in the first 3 seconds drives a 14% higher message recall. 🧠

Meta's Robyn and Google's LightweightMMM stand as two open-source advanced MMM platforms. While both tools aim to offer robust modeling capabilities, they differ substantially in philosophy, implementation, and their approach to data optimization.

In this article, we dive into the capabilities and limitations of two leading open-source platforms - Robyn developed by Meta and Google's Lightweight MMM platforms. While both tools promise greater efficiency and precision in marketing analytics using AI, they differ substantially in their technical implementation. We compare key features including model architecture, optimization methods, adstock handling, seasonality, and more. The analysis also covers areas requiring additional expertise when applying these advanced modeling tools.

For marketers looking to upgrade their analytics tech stack with AI-powered measurement solutions, understanding the core strengths and tradeoffs of Robyn versus Lightweight provides valuable guidance on selecting the solution that’s best suited for your business needs. The overview helps highlight the most important factors in evaluating how well each platform can address these needs. As MMM enters a new era of AI innovation, the capabilities mapped out in this comparison serve as an important benchmark for the marketing analytics space right now.

Robyn’s Features

Programming Language and Accessibility

Robyn is built using the R programming language, a choice that arises from traditional econometricians and statisticians who predominantly work in R. However, while R is robust for statistical analysis, it is not as widely used as Python, making it less versatile for broader teams that may be more familiar with Python.  Robyn’s PM at Meta explored a Python wrapper for Robyn, but opted to hold off until all of Robyn’s major features shipped.   

Model Architecture and Run Time

Compared to Google's LightweightMMM, Robyn takes considerably longer to run models. One of Robyn's distinguishing features is its complex model architecture, employing a multi-stage approach. A Ridge regression model is run, followed by the optimization of hyperparameters through Nevergrad's evolutionary algorithm. Additionally, Robyn’s attention to Ad Stock effects is more robust.  This complexity results in a longer model run time, but the trade-off is a more nuanced understanding of marketing dynamics.

Advanced Optimization Metrics

For hyperparameter optimization, Robyn goes beyond MAPE (Mean Absolute Percentage Error) and employs its own metric, Decomp.RSSD. It stands for “Decomposition Root Sum of Squared Distance.”  Decomp.RSSD minimizes the distance between share of spend and share of effect, to get rid of the extreme disproportionate results based between media channels. This  makes the model’s recommendations more plausible, avoiding suggestions of drastic changes in ad spend that may not be feasible or effective in practice.

Detailed Adstock and Saturation Models

Robyn also outshines in its handling of adstocks and saturation effects. It uses a robust variety of models for these factors and dedicates significant computational power to optimize them. In comparison to alternatives like Google's LightweightMMM, Robyn offers a more comprehensive approach by building around 10,000 models to estimate the parameters for adstocks and saturation effects.  

Seasonality and Domain Knowledge

Seasonality is another area where Robyn excels. It utilizes Prophet, a Bayesian open-source library, to manage seasonality effectively by accounting for annual seasons, holidays, and other events specific to the country. Moreover, Robyn allows for the inclusion of domain knowledge as a specific feature and optimization target, which is particularly valuable for businesses with specialized market conditions or product offerings.

Decision-making Tools

Robyn provides a rich set of decision-making tools, including waterfall charts and metrics like the share of effect versus the share of spend. These add an extra layer of insight, aiding marketers in comprehending how various channels contribute to overall performance.

Time to Run the Model

Despite its more advanced features, Robyn doesn't require an extended period for project completion compared to other solutions. Generally, defining the business scope, data collection and cleaning, and presenting the results will take approximately 8 to 14 weeks, which is comparable to Google’s LightweightMMM.

Areas of Consideration

Complexity

While Robyn offers robust and comprehensive solutions, there are a few things marketers should be aware of. For companies that don't have the technical expertise, the learning curve can be steep. The complexities associated with its multi-stage modeling and advanced metrics might not be easily grasped by marketing teams without a strong statistical background.  The reliance on R could be a constraint for teams more familiar with Python. 

Domain Knowledge Requirement

While Robyn provides features for incorporating domain knowledge, these require expertise and manual adjustment. Marketers without a deep understanding of domain-specific intricacies may find it difficult to fully leverage this feature.

Shifting Modeling Windows

Robyn offers a shifting modeling window that allows companies to zoom into more recent data, like the last 3 or 6 months. While this feature is beneficial for incorporating fresh data, it could be problematic for channels with long adstocks or shift effects, leading to significant changes in model parameters that could be misleading.

Google’s Lightweight

Google's LightweightMMM emerges as a modern solution, providing agile and efficient ways for marketers to make data-driven decisions. Here, we dive into the capabilities of this innovative platform and why it's catching the attention of the digital marketing world.

Programming Language

In contrast to Meta's Robyn, which is grounded in R, LightweightMMM is built on Python. Google's embrace of Python makes the LightweightMMM platform more accessible, tapping into a larger pool of developers and data analysts familiar with the language.  On the other hand, it may not appeal to traditional Econometricians and Statisticians that prefer R for Marketing Mix Modeling.

Rapid Model Run Time

Speed is one of the standout features of LightweightMMM. This expeditious computation is due to LightweightMMM's configuration, which incorporates the Bayesian model to natively estimate saturation curves and adstock rates. Additionally, its integration with Numpyro and JAX for differentiation offers a 'lightweight' approach, true to the platform's name.

Advanced Optimization

In terms of optimization, LightweightMMM takes a different route than its counterparts. The model emphasizes accuracy, optimizing based on the MAPE (Mean Absolute Percentage Error). This metric provides insights into the model's average day-to-week accuracy. However, to avoid implausible results, Google employs a Bayesian framework. This method sets "priors" that act as safeguards against unlikely outcomes for marketing coefficients. Moreover, Google's media budget optimizer generally suggests only up to a 20% change in spend, ensuring recommendations remain realistic and actionable.

Handling Adstocks and Diminishing Returns

Google's approach to adstocks and diminishing returns is modular and scalable.  LightweightMMM provides three main models related to media spend. The default “adstock” model captures the lagged effects of ad spend on sales. For businesses with larger ad budgets that need to account for the diminishing performance as ad spend increases, the "hill_adstock" model is available. Google's method, while simpler than Robyn's multi-model approach, is effective, especially for smaller businesses that may not experience significant saturation effects.

Seasonality

LightweightMMM takes a native approach to seasonality, using a sinusoidal parameter for repeating patterns. The inclusion of an Intercept, trend, and error terms ensures a comprehensive understanding of seasonal variations. Although Google's approach is somewhat different from Robyn's use of the Prophet library, it still addresses the fundamental need for seasonality in modeling, allowing for more strategic marketing decisions during peak periods.

Geo-level Support

Arguably, one of the most distinguishing features of LightweightMMM is its native support for geo regions. When provided with data from multiple geographic areas, the model outputs individual charts for each region, coupled with overarching accuracy and performance metrics. Such granularity not only enhances the model's accuracy but also allows businesses to make more tailored decisions based on geographic nuances.

Domain Knowledge and Adaptability

While LightweightMMM doesn't come with out-of-the-box features for domain knowledge like Robyn, it offers flexibility within its Bayesian framework. Skilled data scientists can modify the platform to accommodate their unique domain knowledge, fine-tuning the model for specific market conditions or business nuances.

Areas of Consideration

Limited Documentation

Google's project is still relatively underdeveloped compared to Meta’s Robyn. It lacks extensive supporting documentation and guides. This could make the initial setup and ongoing management of the platform more challenging for users who are new to MMM.

Less Advanced Optimization

LightweightMMM optimizes based on MAPE alone, which is a measure of accuracy. It doesn't offer advanced metrics like Decomp.RSSD or MAPE.LIFT (for experiments) that Robyn uses. While MAPE is a good measure of average model performance, it may not account for some nuances in spend allocations, potentially making the model less actionable.

Handling Adstock and Saturation

While LightweightMMM offers three main models related to media spend, it doesn't provide as many options for adstock and saturation as Robyn. For businesses with more complex needs, LightweightMMM might not offer the kind of specificity required to accurately model the diminishing returns of advertising spend.

Seasonality Handling: Fixed Assumptions

Although LightweightMMM accounts for seasonality natively, it uses a sinusoidal parameter with a repeating pattern. This approach assumes fixed seasonality effects, which might not be appropriate for all businesses, especially those with variable seasonal trends.

Lack of Domain Knowledge Integration

LightweightMMM doesn’t natively support the integration of domain-specific knowledge in the way that Robyn does. While Bayesian frameworks like LightweightMMM allow for the use of priors, this requires manual intervention and a high level of expertise, making it less accessible for less technical users.

No Feature for Recent Performance Understanding

LightweightMMM doesn’t offer the same flexibility as Robyn in terms of modeling window adjustments to focus on more recent performance data. This can be problematic for businesses with fast-changing environments or those that want to incorporate the most current data, provided that the datasets are large enough, into their models.

Each tool has its distinct advantages and drawbacks. Marketers should consider their specific needs, including the expertise available, the time they can allocate, and the complexity of their marketing channels, before choosing between Robyn and LightweightMMM.