This section refers to a deprecated version of the product. The new version is FE&R. To access FE&R, contact your CSM.
📘 To learn more, read the FE&R documentation.
LogoLogo
PlatformsPricingRessources
  • User documentation
  • Onboarding
  • Help Center
  • Release Notes
  • Flagship - Deprecated
  • Feature Experimentation & Rollout (ex-Flagship) is evolving!
  • GETTING STARTED
    • 👉Managing multiple environments
    • Using the trial version of Flagship
  • First steps with Flagship
    • Quick start guide
    • Glossary
  • Implementation
    • Sending transactions from the AB Tasty Shopify app directly to Feature Experimentation & Rollouts
  • Integrations
    • Heap Analytics integration (Push)
    • Tealium AudienceStream (receive audiences)
    • FullStory integration (segment exports)
    • Heap Analytics integration (Pull)
    • Google Analytics integration (pull audiences)
    • Segment Integration (receive traits)
    • Mixpanel integration (cohort exports)
    • 👉Retrieving your third-party tools’ audiences in AB Tasty - Feature Experimentation & Rollouts
    • Zapier integration
    • Segment integration
  • Steps configuration
    • 👉Configuring Sequential Testing Alerts
    • 👉Configuring your Flags
    • 👉Using the scheduler
    • 🛠️[Troubleshooting] How to target a large number of users at the same time?
    • 👉Configuring KPIs
    • 👉Using the automatic rollback option
    • 👉Targeting configuration
    • 👉Dynamic allocation
    • 👉Traffic allocation
  • Team
    • Access Rights, Teams & User Management
    • 👉Defining rights per project
  • DEMO
    • AB Tasty - Feature Experimentation & Rollouts Demo - How to use it
  • Navigating the interface
    • 👉Archiving use cases from the dashboard
    • 👉Flags page
    • 👉Running a search on the dashboard
    • Navigating the Flagship interface
  • REPORTING
    • 👉Verifying your hit setup
    • 👉Exporting reporting data
    • Understanding the "Chances to win" indicator
    • 🛠️[Troubleshooting] How can I know my test is reliable and my data significant enough to be analyzed?
    • Reporting - A/B Test
    • 👉Using the reporting filters
  • API keys & Settings
    • 👉Acting on your account remotely
    • 👉Using Experience Continuity
    • visitor experiment option
  • FEATURES SETUP
    • 👉Bucket allocation
  • SDKs integration
    • 👉Managing visitor consent
    • 👉Understanding the use of SDKs
  • FAQ
    • Can I make a comparison for each reporting?
    • Can I use Flagship even if my SDK stack is not available?
  • Platform integration
    • 👉Webhooks page
  • Decision API
    • Decision API for non-techie users
  • Account & Profile
    • 👉Configuring account restrictions with MFA
    • 👉Configuring a FA on your profile
  • RELEASE NOTES
    • October - Flagship becomes Feature Experimentation & Rollouts
    • February - Release Notes
    • 📅January - Release Notes
    • 🎉December - Release Notes 🎉
    • 🦃November - Release Notes
    • September Release Notes 🎨
    • June Release Notes 🐞
    • 🍸May Release Notes ☀️
    • Flagship Release Notes April 🐇
    • Flagship February release notes 🏂
    • Flagship January release notes 🎉
    • Flagship November release notes 🦃
    • Flagship October Release Notes 🎃
    • Flagship September Release note 🎒
    • Flagship August Release Notes 🐬
    • Flagship Release Notes July ☀️
    • Flagship Release notes June 🌻
    • Flagship Spring Release May 🌸
    • Flagship Release Notes: Fall
  • Use cases
    • 👉Duplicating a winning variation
    • 👉Configuring a Feature Toggle/Flag
    • 👉Configuring an A/B Test
    • 👉Configuring a Progressive rollout
    • 👉Configuring a Personalization
  • VIDEO TUTORIALS
    • [Video Tutorial] AB Test
    • [Video Tutorial] Feature Flag
    • [Video Tutorial] Progressive Deployment
Powered by GitBook
LogoLogo

AB Tasty Website

  • Home page AB Tasty
  • Blog
  • Sample size calculator
  • Release note

AB Tasty Plateform

  • Login

© Copyright 2025 AB Tasty, Inc, All rights reserved

On this page
  • 📖 Definition
  • ⚙️ Configuration

Was this helpful?

Edit on GitLab
Export as PDF
  1. Steps configuration

Configuring Sequential Testing Alerts

📖 Definition

Sequential Testing Alerts offers a sensitivity threshold to detect underperforming variations early and trigger alerts or pause experiments. It enables you to make data-driven decisions swiftly and optimize user experiences effectively. Sequential Testing saves time and resources by stopping experiments that are unlikely to yield positive results.

⚙️ Configuration

To enable Sequential Testing alerts on an experiment, apply the following steps:

  1. Access the AB Test configuration

  2. At the bottom of the first step, the Main information step, click on the Advanced Options

  3. You can then configure 2 options:

    1. The type of automation You can either receive an email if your campaign is failing, a kind of “Warning, you’re experiment is going right into the wall” or directly pause the campaign if it’s failing.

    2. The sensitivity threshold you want to experiment on

      You have 3 different level of sensitivity here:

      • High - It’s the safest solution. You’ll be more likely to receive alerts but you’re sure that losing experiments will be detected sooner.

      • Low - It’s the more advanced setup. You’ll receive less alerts, but your losing experiments will take longer to be detected.

      • Balanced - It’s the half-way between the 2 previous detection levels.

You can learn more about detection level on the Sequential Testing Alert article.

Also, more information about Sequential Testing in the FAQ article

Good to know 💡

You can only configure this setup if you selected a Primary Goal at the previous step!

PreviousSegment integrationNextConfiguring your Flags

Last updated 2 days ago

Was this helpful?

👉