# A/B Testing

{% hint style="info" %}
Available from the Growth plan.
{% endhint %}

### Overview

A/B Testing lets you compare two versions of a demo and understand which one performs better based on **real user behavior**.

Instead of guessing or debating internally, you can run controlled experiments and optimize demos for:

* Higher completion rates
* More CTA clicks
* More leads captured
* Stronger buyer intent

This feature is designed specifically for **marketing and growth teams** using demos as a conversion asset.

{% @storylane/embed subdomain="app" linkValue="xos6sibfdhof" url="<https://app.storylane.io/share/xos6sibfdhof>" %}

***

### Who this feature is for

* Product marketing teams
* Growth & CRO teams
* Demand generation and website teams

If demos are part of your acquisition, activation, or lead capture flow, A/B Testing helps you continuously improve performance.

***

### How A/B testing works (conceptually)

An A/B test compares:

* **Variant A** - your original demo
* **Variant B** - another demo (ideally, a modified, duplicated version of the same demo)

Traffic is automatically split between the two variants, and Storylane tracks how each version performs across key metrics.

Both variants receive **real traffic**, so results reflect actual buyer behavior - not simulations.

***

### Best practice: d**uplicate the demo for variant B**

Before creating an A/B test, we suggest duplicating your original demo, making changes, and using it as variant B.

This ensures:

* Identical demo size and layout
* Clean, comparable results
* Only one variable is being tested

> ⚠️ Changing multiple things at once makes results harder to interpret.

**Recommended:**\
Duplicate → change one thing → test.

***

### What you should test

Marketing teams commonly test:

* Lead form vs no lead form
* Short vs long demo
* CTA at the start vs CTA at the end
* Gated vs ungated experience
* Use-case intro vs feature-first intro
* Video + voiceover vs silent demo
* Strong CTA copy vs soft CTA copy

Always test **one variable at a time**.

***

### Understanding Traffic Split

#### Default: 50/50

Best for:

* Most experiments
* Faster, balanced learning

#### When to Adjust the Split

You may want to change the traffic split when:

* Protecting a high-performing demo
* Testing a major structural change
* Running experiments during high-traffic campaigns

**Example:**

* 80% → proven demo
* 20% → experimental version

This balances learning speed with conversion risk.

***

### Understanding Test Results

#### Clear Winner

One variant consistently outperforms the other across key metrics.

#### No Clear Winner

This is still a valuable outcome.\
It tells you that the tested change did **not** significantly impact performance.

Use these learnings to inform your next experiment.

***

### Ending an A/B Test

When you’re ready, you can end the test to stop traffic splitting.

#### If Variant A Wins

* No action required
* Your existing embedded demo link already points to Variant A

#### If Variant B Wins

* Update the demo link wherever it’s embedded\
  (website, landing pages, campaigns, emails)
* This ensures future traffic goes to the better-performing demo

Ending a test is a deliberate decision and cannot be undone.

***

### Tips for Better Experiments

* Test one variable at a time
* Let the test run long enough to collect meaningful data
* Avoid ending tests too early
* Use results to plan the next iteration

A/B Testing works best as a **continuous optimization loop**, not a one-off task.


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://docs.storylane.io/analytics-and-performance/a-b-testing.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
