Skip to main contentTailor AI LogoTailor AI

    Experiments Workflow

    Everything about running experiments in Tailor: from planning to promoting a winner.

    A/B vs. Multi-Variant (A/B/C)

    Most experiments should be a simple A/B test (control vs. one variant). Adding more variants fragments your traffic and slows learning.

    When to Use Multi-Variant

    Only add a third variant when you have enough conversion volume that splitting traffic three ways won't slow results to a crawl. Variants must be meaningfully different (not tiny copy tweaks).

    If you're struggling to get steady conversions with two variants, stick to A/B. Sequential iteration (promote the winner, then test the next idea) is almost always better than testing five variants at once.

    Setting & Changing the Control

    The control variant is the baseline your treatment is measured against. If you want to change the control mid-experiment, the best practice is to stop the experiment, set the new control, and start a fresh test. Changing the control mid-flight contaminates the measurement window and can produce false positives.

    Ramping Safely

    Tailor defaults to a 50/50 traffic split, which works for the vast majority of experiments. Gradual ramping (starting at 10% and increasing) is only needed for very high-volume pages with sensitive outcomes.

    Recommended Approach

    • Launch at 50/50 unless you have very high traffic and a sensitive page
    • Monitor results after launch
    • Deramp immediately if you see a clear negative trend

    Reading Results

    Use this five-point framework to evaluate your experiment:

    • 1Primary goal: did the variant improve meaningfully vs. control on your conversion goal?
    • 2Volume: do you have enough conversion volume to trust the direction?
    • 3Consistency: does the lift hold across days and major segments?
    • 4Diagnostics: are clicks up without downstream lift? That's a warning sign (engagement without conversion).
    • 5Confounds: did campaigns, deploys, or tracking change mid-test?

    "Too Early" / Insufficient Data

    "Too early" means there isn't enough signal on your goal to separate real lift from noise. There's no magic number for how long to wait. It depends on stable conversions per variant over time.

    How to Speed Up Results

    • Increase conversion volume by driving more qualified traffic
    • Reduce variants (A/B instead of A/B/C) to concentrate traffic
    • Temporarily use a higher-frequency proxy goal (e.g. form starts instead of form submits)

    Broken tracking can make results look "too early" indefinitely. If data isn't accumulating, check that your conversion goal is firing correctly on both control and treatment before waiting longer.

    Minimum Traffic Needed

    There's no universal minimum, but the rule of thumb is: you need enough conversions that you're not reading noise. If your page converts at 2% and you get 100 visitors/week, that's about 2 conversions per variant per week. It will be very hard to detect lift at that volume.

    For low-volume pages: stick to A/B (not multi-variant), make bigger and more meaningful changes, and consider a higher-frequency proxy goal to validate direction before committing to a longer test.

    Promoting a Winner

    When you're confident in the results, here's the workflow:

    1

    Confirm lift on primary goal

    Make sure the improvement is on your real conversion goal, not just clicks or engagement metrics.

    2

    Ramp to 100%

    Click "Ramp to 100%" to send all traffic to the winning variant.

    3

    Start the next test

    Use the winner as the new control and test one new hypothesis. Don't pile multiple changes into one test.

    4

    Monitor for regression

    Watch metrics for a few days after promotion. Novelty effects and promo bursts can fake a winner.

    Stopping a Test Safely

    To stop an experiment without losing data:

    • 1Deramp the variant (set allocation to 0%). This stops exposure without deleting anything.
    • 2Confirm the control experience is serving normally.
    • 3Record the final readout: dates, traffic, conversions, and lift.
    • 4Decide: promote the winner, iterate on the variant, or scrap it.

    Deramp vs. Delete

    Always deramp first. Deramping keeps the variant for future reference or re-testing. Only delete if you're certain you won't need it again. Available experiment actions are: ramp, ramp to 100%, deramp, and delete. There is no pause action.

    Scheduling Experiments

    Experiment scheduling (e.g. start Monday, end Friday) is not supported today. The workaround is to QA your experiment ahead of time, then launch and stop it manually. This is a known request and may be added in the future.

    Related Guides