LogoLogo
PlatformsPricingRessources
  • Home
  • Onboarding
  • User documentation
  • Release Notes
  • Help Center
  • User documentation
  • Dashboard
    • Global Experience Dashboard discovery
    • How to filter the Global Experience Dashboard
    • How to customize your Global Experience Dashboard
    • Uplift card
  • Web Experimentation and Personalization
    • Campaign creation and dashboard
      • Types of campaigns
        • Choosing the right type of campaign with Ally, our assistant
      • How to create a campaign
        • Experimentations
          • How to create an A/B Test
          • How to create a Multipage Test
          • How to create a Split Test/Test by Redirection
          • How to create an A/A Test
          • How to create a Patch / Multipage Patch
          • How to create a Multivariate Test
        • Personalizations
          • How to create a Multi-Experience Personalization
          • How to create a Multipage Personalization
          • How to create a Simple Personalization
      • Campaign duplication
        • How to duplicate a campaign
      • How to set-up Main Information step
      • Understanding campaign duration
      • Campaign loading (deferred/ instant)
      • The "Comment" section of the campaign creation flow
      • Campaign statuses
      • How to exclude IP addresses from your campaigns
      • How to use MDE Calculator
      • Campaigns Dashboard
      • Campaign Scheduler
      • Mutually Exclusive Experiments
      • How to use Campaign Prioritization
        • Prioritization of Personalizations
    • Editors and Widget
      • How to create and edit content in the visual editor
      • How to use our Visual Editor - Interactive demo
      • Discovering the Visual Editor
      • Visual editor - History and review of modifications
      • Code Editor
      • Using the Editor Copilot
      • Prompt Inspiration library for Visual Editor Copilot
      • How to create an Action Tracker in the editor
      • How to use redirection option
      • How to create trackers in the editor
      • Widgets
        • Widgets configuration in the visual editor
        • AB Tasty prebuilt Widgets
          • Action Button Widget
          • Banner Widget
          • Before-After Image Widget
          • CSAT Widget
          • Celebrate Widget
          • Countdown widget
          • Drawer Widget
          • Dwell Time Tracking Widget
          • Element Visible Tracking Widget
          • Iframe Click Tracking Widget
          • IFrame Widget
          • Image Pop-in Widget
          • NPS® (Net Promoter Score) Widget
          • Progress Bar Widget
          • Scratch card Widget
          • Scroll Rate Tracking Widget
          • Search & Replace Widget
          • Simple Popin Widget
          • Snowflake Animation Widget
          • Social proof Widget
          • Social Sharing Widget
          • Spotlight Widget
          • Sticky Element Widget
          • Tooltip Widget
          • Video Pop-in Widget
          • Virtual Click Widget
        • Custom widget
          • Custom Widgets: List of propname Fields and Their Specificities
          • Custom widgets: Global usage and information
        • Custom code widget
      • Chrome extension
      • Shadow DOMs & iFrames
        • Understanding Shadow DOM
        • How to edit Shadow DOM in the visual editor
        • Understanding iframes
        • How to edit Iframes in the visual editor
    • Goals step
    • Targeting step
      • How to set up a campaign Targeting
      • How to create a Segment (Who section)
        • Segments Recommendations
        • Inspiration with Engagement Level segmentation
      • How to define Targeted Pages (Where section)
      • How to create a Trigger (How section)
        • Geolocation criterion
        • Device criterion
        • Screen size criterion
        • IP address criterion
        • Weather criterion
        • Landing Page criterion
        • Source & source type criterion
        • Previous Page criterion
        • URL Parameter criterion
        • Minimum Pages viewed criterion
        • Visited Pages criterion
        • Same day visit criterion
        • Browser language criterion
        • Browser criterion
        • Adblocker criterion
        • JS Variable criterion
        • Code-based criterion
        • CSS Selector criterion
        • DataLayer criterion
        • Visitor attributes criterion
      • How to define a Targeting Frequency (When section)
      • How to set-up targeting verifications
      • How to use the replicate targeting option
      • How to use URL operators to build URL patterns
    • Traffic allocation step
      • Campaign flow: Traffic Allocation step
      • Dynamic allocation
      • Even allocation
    • Advanced Options step
      • Sequential Testing Alerts
    • QA step
      • QA Mode & QA Assistant
        • How to use the QA mode
        • How to use the QA Assistant on desktop
        • How to use the QA Assistant on mobile
      • How to use the browser console to QA a campaign
  • Feature Experimentation & Roll-Out
    • How to create server-side A/B tests
    • How to create feature toggles and manage rollouts
    • How to create server-side personalization
    • FE&R Glossary
    • Managing Flags
    • Managing Context Keys
  • Recommendations & Merchandising
    • Recommendations
      • Web recommendation
        • Most popular recommendation filters
        • Variables by placement
        • Placement
        • Products groups
        • Filters
        • Exceptions
        • Options
      • Recommendation Banner
        • Recommendation banner with tag and custom JS
          • Adding Recos Tag to a website
          • Enriching Recommendations Banner HTML to send events to DataLayer through Recos Tag
          • Setting up GTM to send Recos Datalayer Events
          • Connect Tag Assistant from GTM
          • Check that the events arrive correctly in GA
          • Check the display of recommendations
          • Retrieving recos products from a RECO_ID with Recos Tag
        • Deploying a Recommendations Banner Custom Widget
        • Retrieving a Recommendation RECO_ID
        • How to push recommendations with Adobe Campaign
        • How to build recommendations (recommendation builder variant)
      • Mail recommendations
        • How to push recommendations into your Brevo emails
        • Brevo Feed
        • Products groups
        • Filters
        • Products highlights
        • Exceptions
        • Options
        • Parameters
        • Preview
        • Most popular recommendation filters
    • Merchandising
      • Introduction
      • Category
      • Products groups
      • Filters
      • Products highlights
      • Exceptions
      • Result to refine
      • How to deploy a merchandised product list through Algolia
    • Recos & Merch analytics
      • How to track performances
      • Recos Analytics Event format
      • The impact of BigQuery exports in terms of cost
      • Setup analytics retrieve events with Recos Tag
      • Key indicators over the period & Segment
      • Overview data
      • Evolution metrics
    • Recos & Merch API
      • Getting a product list
      • Accessing Recos API (endpoint)
      • Accessing Recommendations API (only specific accounts)
    • Recos & Merch account management
      • How to manage users in Recommandation & Merchandising module
      • How to manage roles
      • How to manage synchronisations
      • How to manage invoices
    • Security and GDPR
      • Procedure for Reporting Personal Data Breaches
      • Backup Policy
      • Security Configuration of Our Servers
      • PSSI
    • Recos & Merch Algorythms
      • How to manage algorithms
      • How to create your algorithms
      • Better understanding
      • Default algorithms
      • Data used to train Algorithms
    • Recos & Merch integrations
      • Integrations with Recos & Merch
      • Brevo Integration
      • Analytic integrations
        • Google Analytics integration
        • Matomo integration
        • Random data integration
      • Recos & Merch CMS integration
        • PrestaShop integration
        • Shopify Integration
        • Custom integration
  • Emotions AI
    • Understand EmotionsAI
      • EmotionsAI in a nutshell
      • Leverage EmotionsAI to get more value from A/B tests
      • Leverage EmotionsAI to get high-potential A/B test ideas
    • First steps with EmotionsAI
      • Onboarding
      • EmotionsAI functioning
      • Navigating the interface
      • FAQ about EmotionsAI
    • Using EmotionsAI
      • How to activate EmotionsAI features
      • EmotionsAI Website Mapping configuration
      • EmotionsAI Journey Analysis
      • EmotionsAI Page Analysis
      • Target EmotionsAI segments
      • QA a campaign on Desktop with EmotionsAI Chrome Extension
      • Use EmotionsAI targeting on hard-coded personalizations
      • Usage of EmotionsAI in some Push integrations
    • Read EmotionsAI campaigns’ report
  • Library
    • How to set-up your libraries
    • Widget library
      • Creating and managing Widgets
      • How to create a preset Widget
      • How to create a Custom Widget
        • How to configure Custom Widgets Custom Forms
      • How to declare your Widget Theme
    • Assets Library
    • Trackers library
      • Action Trackers
        • How to create and manage Action Trackers
      • How to create a Custom Tracker with custom JavaScript code
      • How to create a Custom Tracker via the DataLayer
    • Creating and managing Segments
      • List of Segment criteria
        • CSAT criterion (Customer Satisfaction)
        • NPS® criterion (Customer Satisfaction)
        • Engagement Level criterion
        • EmotionsAI criterion
        • Content Interest criterion
        • Page interest criterion
        • Device criterion
        • Cookie criterion
        • New/Returning visitors criterion
        • Campaign Exposure criterion
        • Number of sessions criterion
        • Day(s) Since Last Session criterion
        • Day(s) Since First Session criterion
        • Action Tracker criterion
        • Cart Abandonment criterion
        • Last Purchase criterion
        • Purchase Frequency criterion
        • Geolocation criterion
        • DMP, CDP, rd party tool criterion
      • Segments Library - Metrics
    • Creating and managing Triggers
      • List of Trigger criteria
    • Creating and managing Saved Pages
  • Integrations
    • Integrations general information
    • Push integrations
      • Adobe Analytics
      • Air
      • Amplitude (analytics browser)
      • Amplitude
      • AT Internet Smart Tag
      • Contentsquare
      • Google Analytics (Audience creation)
      • Google Analytics (UA)
      • Google Analytics (event only)
      • FullStory
      • Heap Analytics
      • Hubspot
      • Matomo
      • Mixpanel
      • Piano Analytics
      • SalesForce Marketing Cloud (Push)
      • Segment
      • Tealium
    • Pull integrations
      • Air
      • Commander Act
      • EmotionsAI
      • Fullstory
      • Google Analytics
      • Heap Analytics
      • Mixpanel
      • mParticle
      • Piano Analytics (Pull)
      • SalesForce Marketing Cloud (Pull)
      • Segment
      • Tealium : Import Audiences
      • Weborama
    • Other integrations
      • Productivity
        • Microsoft Dynamics Commerce
        • Setting up Microsoft Clarity Integration with AB Tasty
        • Didomi
        • How to Set Up the AB Tasty Connector with Notion
        • How to connect AB Tasty to Google Sheets
        • Connect your Slack Workspace to AB Tasty
      • Data Warehouse
        • Data Warehouse integrations: General information
        • Big Query - Daily exports from AB Tasty to Big Query
        • Snowflake - Daily exports from AB Tasty to Snowflake
        • Redshift - Daily exports from AB Tasty to Redshift
    • Custom integrations
      • Universal Data Connector
      • Custom Integration Connector with a rd party tool (Push data)
      • AB Tasty public API
  • Reporting & Performances
    • Data Explorer
      • How to analyze page traffic via Data Explorer
      • How to get metrics breakdown by multiple dimensions via Data Explorer
      • How to get a list of recent hits via Data Explorer
      • Data Explorer - List of Metrics and Dimensions
    • ROI dashboard
    • Performance Center
      • List of Performance Guidelines
    • Reporting
      • Campaign reporting
        • Reporting Copilot
        • Analysis Copilot
      • Data & reports: generalities and definitions
      • AB Tasty reports Metrics
        • Live hits of the reporting
        • Metrics based on tracking widgets
        • Navigation metrics
        • Metrics based on pageviews
        • Metrics based on clicks
        • Metrics based on transactions
      • Reporting Filters
        • General Reporting filters
        • EmotionsAI Reporting filter templates
      • Using the NPS®️ report
        • How does Feedback Analysis Copilot work
      • Sample ratio mismatch
      • Refresh reporting data using Refresh On Demand
      • Data export from the reporting
      • Frequentist Analysis mode
      • Understand the statistics behind your Reports
        • Statistics for the reporting
        • Statistical metrics
        • Avoiding Pitfalls in AOV Analysis for Conversion Rate optimization
        • Conversion mechanisms & concepts
        • Reporting Readiness
  • Account
    • Tag integration
      • All About Tags
      • How to implement the Generic Tag
        • AB Tasty hosted tag implementation
        • Tag Domain Delegation
          • Tag domain delegation implementation
      • How-to QA the Generic Tag
      • How to implement the Generic Tag via Google Tag Manager
      • How-to join the Next tag program
      • AB Tasty integration with Shopify
        • Understanding Shopify App features
        • How to implement the AB Tasty tag via Shopify
        • How to set-up Shopify Custom Pixel App
      • AB Tasty tag compilation
      • How the AB Tasty tag is designed to handle Single Page Apps (SPA)
    • Technical implementation
      • How to choose your cookies deposit method
      • How to declare my Account Domain?
      • JavaScript in AB Tasty
        • How to configure JavaScript
        • Javascript files execution
        • Campaign JavaScript Execution
      • How to configure jQuery loading
      • How to integrate Product hits
      • Product Hits integration principles
        • How to set-up segment criteria "Content Interest" and "Cart Abandonment"
    • Transaction Tag integration
      • How to implement the Transaction tag
      • How to create a Transaction Tracker via DataLayer
      • How to use the Transaction Tag Generator
      • How to QA the Transaction tag
    • Performance and security
      • How to manage visitor identity
      • How to deactivate AB Tasty
      • Consent policy - cookies, storage and privacy
    • Account management
      • The organization page
      • Subscription page
      • How to manage users?
Powered by GitBook
LogoLogo

AB Tasty Website

  • Home page AB Tasty
  • Blog
  • Sample size calculator
  • Release note

AB Tasty Plateform

  • Login

© Copyright 2025 AB Tasty, Inc, All rights reserved

On this page
  • What Is AOV?
  • Common Mistake: Outliers in A/B Tests
  • Example: The Impact of a High-Value Order
  • Why AOV can be misleading
  • Impact on Related Metrics
  • How to Analyze AOV Correctly
  • Conclusion
  • Related articles:

Was this helpful?

Edit on GitLab
Export as PDF
  1. Reporting & Performances
  2. Reporting
  3. Understand the statistics behind your Reports

Avoiding Pitfalls in AOV Analysis for Conversion Rate optimization

When optimizing your conversion rate (CRO), it’s easy to rely on Average Order Value (AOV) as a key metric. However, AOV can be misleading and often results in incorrect conclusions if misinterpreted. Let's explore why this happens and how to use AOV effectively.

What Is AOV?

Average Order Value (AOV) is calculated by dividing the total value of orders by the number of orders. It’s a convenient metric to track how much customers are spending on average. Many CRO professionals use it to compare the performance of different A/B test variations. Unfortunately, focusing on AOV alone can lead to inaccurate insights.

Common Mistake: Outliers in A/B Tests

Imagine running an A/B test to compare two versions of your website. If just one customer makes an unusually large purchase, the AOV for that variation could spike significantly. This single outlier can make it seem like the variation is outperforming the other, even though it might not be true in a broader context. This issue is exacerbated in small sample sizes, but even large datasets are not immune.

Example: The Impact of a High-Value Order

Consider an A/B test where two subgroups are being compared:

  • Variation A has an AOV of €56.6 with 5 orders of value 50, 51,30, 75, and €77

  • Variation B has an AOV of €56.8 with 5 orders of value 52, 50, 40, 62 and €80, which is a minor difference.

At first glance, this small difference seems insignificant. Now, imagine that a new customer places an order worth €202 in Variation B. This dramatically increases the AOV of Variation B to €81, with 6 orders of value 52, 50, 40, 62, 80 and €202.

This spike could falsely suggest that Variation B is significantly better, even though it is due to just one high-value purchase. Similarly, the opposite can also occur: an outlier could make it appear that a variation is underperforming when there is no real effect, or even obscure a winning variation, making it look like a loser. This kind of misleading result is more likely to occur with smaller datasets, but even larger datasets can still be influenced by similar outliers. It's important to recognize that these effects can lead to incorrect conclusions in both directions, emphasizing the need for careful analysis.

Why AOV can be misleading

We will break down why relying on Average Order Value (AOV) can be misleading in conversion rate optimization (CRO). There are two key factors that impact the reliability of AOV as a metric:

  1. The number of customers: This is a factor of measurement reliability. Generally, we expect that as the number of customers increases, our metrics should become more stable.

  2. The magnitude of the maximum value: This is a factor of measurement imprecision. As the number of customers increases, the highest values in the dataset are also likely to increase, which can introduce instability into the metric.

The "law of large numbers" is often mentioned in this context as a reason to trust AOV when the number of customers is large enough. However, it is crucial to understand that increasing the number of customers does not always solve the problem. In fact, as the customer base grows, the maximum values in order size are also likely to grow, which means that these two factors effectively counterbalance each other. As a result, the AOV can remain unstable even with a larger dataset.

To illustrate this, let’s look at data collected from an e-commerce site. In the graph showing basket values:

  • The horizontal axis represents the number of basket values collected.

  • The vertical axis represents the maximum value found in the list of collected basket values.

From the graph, we see that as more data is collected, the maximum values tend to increase. This behavior can affect the stability of AOV. Next, let’s consider the evolution of the average value as we collect more data:

The horizontal axis represents the number of baskets collected, and the vertical axis represents the average value of the baskets collected.

At first glance, it may seem that the average value stabilizes quickly. However, upon closer inspection, it takes 20,000 customers to reach what seems like stabilization. Achieving 20,000 customers actually requires significantly more visitors due to conversion rates. For instance, with a typical conversion rate of 5%, you would need 400,000 visitors (calculated as 400,000 * 0.05 = 20,000). Therefore, to conduct an A/B test effectively, you would require a traffic volume equivalent to 800,000 visitors.

Even when it appears that AOV is stabilizing, this stabilization is often deceptive. If we zoom in on the data beyond 20,000 customers, we notice fluctuations of -€1 to +€1 as more data is collected.

This means that, even with a large number of data points, AOV can still vary significantly—by +/- €2 in total. In an A/B test context, this implies that a seemingly stable AOV could indicate a false difference between two variations, even when no real effect exists.

In other words, an A/A test—where there is no difference between the variations—could still indicate a difference in AOV of +/- €2 simply due to fluctuations. This makes it impossible to confidently measure any effect smaller than +/- €2 just by observing AOV.

Impact on Related Metrics

This problem extends beyond AOV:

  • Revenue per Visitor (RPV): Since it divides the total order value by the number of visitors, RPV can also show large, misleading jumps, especially influenced by big spenders.

  • Total Revenue: Summing all cart values can be misleading, even without explicitly averaging. Comparing these large figures can lead to similar effects as averaging, particularly when the number of visitors to each variation is similar. Rare, high-value orders may not be evenly distributed across variations, creating an illusion of significant differences.

How to Analyze AOV Correctly

To make informed decisions, use statistical tools like the Mann-Whitney U test, specifically designed for data like AOV. This test helps you determine whether a difference in AOV between two groups is statistically significant by comparing the distributions of order values in each group, rather than relying solely on raw AOV numbers, which can be misleading due to outliers. Unlike a simple average comparison, the Mann-Whitney U test evaluates whether the overall ranking of values differs significantly between the two groups, making it more robust in handling variations caused by extreme values.

It’s not uncommon for the results from the Mann-Whitney test to contradict the simple AOV measurements—for example, indicating a loss even when the AOV appears to show a gain. This is why statistical validation is crucial.

Conclusion

While AOV is a useful metric, relying solely on it in A/B testing can lead to erroneous conclusions due to its susceptibility to outliers and inherent fluctuations. Always back AOV analysis with appropriate statistical testing to ensure your CRO decisions are based on stable, meaningful data.

Related articles:

PreviousStatistical metricsNextConversion mechanisms & concepts

Last updated 11 days ago

Was this helpful?

Statistical metrics
Reporting Copilot
Statistics for the reporting