Analyse & Testing

A/B Testing: Definition, Bedeutung & Beispiele im Direktmarketing

A/B Testing A/B testing is a statistical testing method in which two variants of a marketing piece (A and B) are sent to comparable, randomly split recipient groups to determine which variant achieves a higher response rate, conversion rate, or better ROI.

Auf einen Blick

Begriff:A/B Testing
Kategorie:Analyse & Testing
Englisch:A/B Testing / Split Test
Synonyme:Split Test, Split-Run Test, Control Group Test

What is A/B Testing? — A Clear Definition

A/B testing is a data-driven testing method in which two variants of a marketing piece are sent to comparable, randomly split recipient groups. Variant A is the "control" (the proven original), while variant B is the "variation" (the modified version with exactly one change). By comparing results — response rate, conversion rate, or ROAS — you can determine on a data basis which variant performs better. The key principle: only a single variable is changed, so that the effect of the change can be causally attributed.

The method has deep roots. British statistician Ronald A. Fisher developed the foundations of modern experimental design in the 1920s at the Rothamsted Experimental Station — including randomization, analysis of variance, and the now-standard 5 percent significance level. Almost simultaneously, American advertising pioneer Claude Hopkins described systematic split testing in advertising for the first time in his book "Scientific Advertising" (1923): he sent different versions of direct mailings to comparable groups and evaluated responses through coded coupon codes. Hopkins' principle holds true to this day: No campaign without a test.

In direct marketing, A/B testing has been the standard method for campaign optimization since the 1960s. The annual CMC Print Mailing Study by Deutsche Post — the largest benchmark study for direct mail in the DACH region (Germany, Austria, Switzerland) — is fundamentally based on A/B tests: every year, over 40 online retailers test different campaign variants on more than one million recipients to gain measurable insights about coupons, formats, response boosters, and design.

6.1%
CVR coupon card — best response booster (CMC 2023)
+58%
Short-term CVR uplift through quantity limitation (CMC 2024)
+25%
Letter vs. postcard CVR advantage (CMC 2020)
1,011%
ROAS of optimized print mailings (CMC 2025)

What Can Be Tested in Direct Mail?

The range of testable elements in direct mail is broad — and not all of them have the same impact on results. The biggest levers are the target audience (which recipients are contacted) and the offer (type and value of the coupon or discount). Only then do format, design, and details such as envelope design or response boosters follow. Those with limited test budgets should therefore start with the most impactful variables.

The CMC studies confirm this prioritization impressively. The CMC Study 2024 tested five different coupon mechanics and found massive differences: the unlimited coupon achieved the highest long-term CVR of 5.4 percent, while time-limited and quantity-limited variants were stronger in the short term (up to +58 percent in the first week) but dropped significantly to just 3.6 percent CVR over the full observation period. By contrast, the CMC Study 2025 showed that envelope design in the B2C segment had no significant impact on conversion — all four tested variants performed nearly identically.

This insight is critical for test planning: you should focus your testing resources on the variables that have the largest expected effect — rather than on design details that produce barely measurable differences. The CMC Study 2022 underscores this: percentage-based coupons achieved 23 percent higher CVR than fixed-amount coupons, and high-value coupons outperformed low-value ones by 61 percent. Such differences in the offer are many times larger than design variations.

Target Audience & Segmentation (Priority 1)

Which recipients are contacted? RFM-based selection can increase CVR by 3.4x according to CMC 2025 — the single biggest lever.

Offer & Incentive (Priority 2)

Type, value, and mechanics of the coupon. CMC 2022 shows: percentage coupons +23% CVR, high-value +61% — massive differences.

Format & Response Boosters (Priority 3)

Letter vs. postcard, coupon card vs. QR code vs. scratch field. Physical inserts (coupon card: 6.1% CVR) outperform gamified elements.

Design & Envelope (Priority 4)

Layout, images, colors, envelope design. CMC 2025 shows: no measurable impact in B2C — content matters more than packaging.

Real-World Examples from the CMC Studies

The CMC Print Mailing Studies deliver the largest A/B tests in direct mail in the German-speaking region year after year. Since 2018, over 40 online retailers have sent more than one million print mailings to existing customers annually, with each edition focusing on a specific A/B test topic. The results set the direction for the entire industry.

The CMC Study 2023 tested four response boosters against each other: an enclosed coupon card, a printed coupon code, a QR code, and a scratch field. The result was clear-cut: the physical coupon card achieved a 6.1 percent CVR — the best value — 33 percent more than the scratch field (4.6 percent). The printed coupon code came in close behind at 5.5 percent. Overall, roughly 20 percent of mailing recipients visited the advertised online shop, with over 70 percent of visits occurring within the first two weeks. The 2020 study tested formats: the classic advertising letter outperformed maxi postcards and self-mailers by approximately 25 percent in CVR and achieved 42 percent higher ROAS than the self-mailer.

A particularly revealing result came from the CMC Study 2024 on coupon limitation: time-limited and quantity-limited coupons drove short-term responses massively upward (+38 percent for a 3-week deadline, +58 percent for quantity limitation) but lost out to the unlimited coupon in the long run. For short-term sales promotions, limitations are therefore effective, but for long-term existing customer reactivation they are counterproductive. This result would not have been detectable without systematic A/B testing — gut feeling would likely have favored the limited variant.

A/B Test Results from the CMC Studies (2020–2025)

Swipe to see more
Year / FocusTested VariantsKey Finding
2025: Envelope Design
4 designs (plain to branded)
B2C: No significant difference
2024: Coupon Limitation
5 variants (unlimited to quantity-limited)
Unlimited wins long-term (5.4% CVR)
2023: Response Boosters
4 variants (card, code, QR, scratch)
Coupon card wins (6.1% CVR, +33%)
2022: Coupon Mechanics
Percentage vs. fixed amount, high vs. low
Percentage +23% CVR, high-value +61% CVR
2020: Format
Letter vs. postcard vs. self-mailer
Letter +25% CVR, +42% ROAS vs. self-mailer
Alternative mobile view:
Year / Focus:2025: Envelope Design
Tested Variants:4 designs (plain to branded)
Key Finding:B2C: No significant difference
Year / Focus:2024: Coupon Limitation
Tested Variants:5 variants (unlimited to quantity-limited)
Key Finding:Unlimited wins long-term (5.4% CVR)
Year / Focus:2023: Response Boosters
Tested Variants:4 variants (card, code, QR, scratch)
Key Finding:Coupon card wins (6.1% CVR, +33%)
Year / Focus:2022: Coupon Mechanics
Tested Variants:Percentage vs. fixed amount, high vs. low
Key Finding:Percentage +23% CVR, high-value +61% CVR
Year / Focus:2020: Format
Tested Variants:Letter vs. postcard vs. self-mailer
Key Finding:Letter +25% CVR, +42% ROAS vs. self-mailer

Methodology — How an A/B Test Works in Direct Mail

A methodologically sound A/B test in direct mail follows a clear process. It begins with the hypothesis: "A 15% coupon with a 3-week deadline achieves a higher CVR than one without a deadline." Next, the recipient list is randomly divided into two equally sized groups — the control group receives the proven mailing, the test group receives the variant with exactly one change. All other factors (timing, target audience, remaining design) stay identical. Both variants are sent on the same day.

The critical question is the sample size: each test variant needs enough recipients to deliver statistically reliable results. The rule of thumb is: the sample should be large enough to generate at least 100 responses per variant. At an expected response rate of 2 percent, that means at least 5,000 recipients per variant — at 1 percent, at least 10,000. The target confidence level is typically 95 percent (p < 0.05), meaning the probability of a chance result is below 5 percent.

Particularly important — and different from digital tests — is the observation window: the CMC studies show that with print mailings, roughly half of all orders only come in from week five onward after dispatch. Anyone who evaluates their test after two weeks captures only a fraction of the effect and risks making wrong decisions. A minimum observation period of six to eight weeks after dispatch is recommended. Only then can the winning variant be identified and scaled up as the new "control" — the next test then tests a new element against this new benchmark.

A/B Testing: Direct Mail vs. Digital

Swipe to see more
CharacteristicDirect MailDigital (Web / Email)
Results available
4–8 weeks
Hours to days
Cost per variant
Print + postage (significant)
Near zero
Sample size
Limited by budget
Infinitely scalable
Response rate
Typically 3–8% (existing customers)
Typically 1–3% (email CTR)
Tracking
Coupon codes, QR, PURLs
Automatic (clicks, views)
Iteration
1–2 tests per quarter
Dozens per week possible
Alternative mobile view:
Characteristic:Results available
Direct Mail:4–8 weeks
Digital (Web / Email):Hours to days
Characteristic:Cost per variant
Direct Mail:Print + postage (significant)
Digital (Web / Email):Near zero
Characteristic:Sample size
Direct Mail:Limited by budget
Digital (Web / Email):Infinitely scalable
Characteristic:Response rate
Direct Mail:Typically 3–8% (existing customers)
Digital (Web / Email):Typically 1–3% (email CTR)
Characteristic:Tracking
Direct Mail:Coupon codes, QR, PURLs
Digital (Web / Email):Automatic (clicks, views)
Characteristic:Iteration
Direct Mail:1–2 tests per quarter
Digital (Web / Email):Dozens per week possible

Common Mistakes in A/B Testing

Even experienced marketers make methodological errors that can invalidate A/B tests. The most common mistake is changing multiple variables at once — for example, altering coupon value, headline, and design in a single variant. In that case, it is no longer possible to determine which change influenced the result. The "One Variable at a Time" (OVAT) principle is therefore the most important methodological rule.

An equally widespread mistake is a sample size that is too small. Testing with only 500 recipients per variant at a 2 percent response rate yields just 10 responses — far too few for statistically significant conclusions. The observed difference could be pure chance. Premature evaluation is another typical error: with print mailings, half of all orders come in only after weeks. Deciding after two weeks may lead to the wrong choice. Other common mistakes include unequal test groups (not randomly split), missing tracking (no individual codes), and adopting results without significance testing.

A/B Testing with AutoLetter

AutoLetter makes A/B testing in direct mail simple and accessible. Instead of coordinating elaborate test designs with traditional mail houses, businesses can configure their campaign variants online and have them sent to randomly split recipient groups — with integrated response tracking for clean performance measurement. This enables data-driven validation of hypotheses about offers, design, or personalization, without having to manually coordinate the testing process.

Especially for businesses that want to systematically optimize their direct mail, AutoLetter offers a decisive advantage: shorter cycles between test and rollout, transparent cost overview per variant, and measurable results as the foundation for the next campaign. Every test makes the following campaign better — fully in line with Claude Hopkins' principle: No campaign without a test.

Optimize your direct mail with data

With AutoLetter, you can easily test different campaign variants with measurable results — for higher response rates and optimal ROI.

Try it for free

Frequently Asked Questions About A/B Testing

5 Fragen beantwortet

The rule of thumb is: at least 100 responses per variant. At an expected response rate of 2 percent, you need at least 5,000 recipients per test variant; at 1 percent, at least 10,000. The target confidence level is typically 95 percent. Industry experts recommend 5,000–10,000 recipients per test cell as a benchmark.

Significantly longer than digital channels. The CMC studies show that roughly half of all orders come in only from week five onward after dispatch. A minimum observation period of 6–8 weeks is therefore recommended. Including preparation (design, printing, list creation) and evaluation, you should plan for 8–12 weeks in total.

The test variables with the greatest impact first: 1. Target audience/selection — has the strongest effect on results (CMC 2025: 3.4x CVR through RFM targeting). 2. Offer/coupon — type and value of the incentive (CMC 2022: +61% CVR for high-value coupons). 3. Format — letter vs. postcard (CMC 2020: +25% CVR for letters). Design details such as envelope design have no measurable impact in B2C according to CMC 2025.

In an A/B test, only one variable is changed (e.g., coupon value) while everything else remains identical — simple, clearly interpretable, and with moderate sample size requirements (5,000–10,000 per variant). In a multivariate test, multiple variables are tested simultaneously (e.g., coupon, headline, and design), which creates many combinations and requires very large sample sizes (5,000–10,000 per combination). For most direct mail campaigns, the A/B test is the more practical method.

The annual CMC Print Mailing Study by Deutsche Post is the largest benchmark study for direct mail in the DACH region. Key A/B test results: coupon card beats scratch field by 33% CVR (2023), unlimited coupons win long-term over limited ones (2024), letter outperforms postcard by 25% CVR (2020), percentage coupons beat fixed-amount coupons by 23% CVR (2022), and envelope design has no significant impact in B2C (2025).

Stay in the Loop

Receive weekly insights on direct marketing, automation, and successful advertising campaigns.

Free and cancellable anytimePrivacy

GDPR compliant
Over 5,000 subscribers