A/B Testing: Definition, Bedeutung & Beispiele im Direktmarketing
A/B Testing A/B testing is a statistical testing method in which two variants of a marketing piece (A and B) are sent to comparable, randomly split recipient groups to determine which variant achieves a higher response rate, conversion rate, or better ROI.
Auf einen Blick
What is A/B Testing? — A Clear Definition
A/B testing is a data-driven testing method in which two variants of a marketing piece are sent to comparable, randomly split recipient groups. Variant A is the "control" (the proven original), while variant B is the "variation" (the modified version with exactly one change). By comparing results — response rate, conversion rate, or ROAS — you can determine on a data basis which variant performs better. The key principle: only a single variable is changed, so that the effect of the change can be causally attributed.
The method has deep roots. British statistician Ronald A. Fisher developed the foundations of modern experimental design in the 1920s at the Rothamsted Experimental Station — including randomization, analysis of variance, and the now-standard 5 percent significance level. Almost simultaneously, American advertising pioneer Claude Hopkins described systematic split testing in advertising for the first time in his book "Scientific Advertising" (1923): he sent different versions of direct mailings to comparable groups and evaluated responses through coded coupon codes. Hopkins' principle holds true to this day: No campaign without a test.
In direct marketing, A/B testing has been the standard method for campaign optimization since the 1960s. The annual CMC Print Mailing Study by Deutsche Post — the largest benchmark study for direct mail in the DACH region (Germany, Austria, Switzerland) — is fundamentally based on A/B tests: every year, over 40 online retailers test different campaign variants on more than one million recipients to gain measurable insights about coupons, formats, response boosters, and design.
What Can Be Tested in Direct Mail?
The range of testable elements in direct mail is broad — and not all of them have the same impact on results. The biggest levers are the target audience (which recipients are contacted) and the offer (type and value of the coupon or discount). Only then do format, design, and details such as envelope design or response boosters follow. Those with limited test budgets should therefore start with the most impactful variables.
The CMC studies confirm this prioritization impressively. The CMC Study 2024 tested five different coupon mechanics and found massive differences: the unlimited coupon achieved the highest long-term CVR of 5.4 percent, while time-limited and quantity-limited variants were stronger in the short term (up to +58 percent in the first week) but dropped significantly to just 3.6 percent CVR over the full observation period. By contrast, the CMC Study 2025 showed that envelope design in the B2C segment had no significant impact on conversion — all four tested variants performed nearly identically.
This insight is critical for test planning: you should focus your testing resources on the variables that have the largest expected effect — rather than on design details that produce barely measurable differences. The CMC Study 2022 underscores this: percentage-based coupons achieved 23 percent higher CVR than fixed-amount coupons, and high-value coupons outperformed low-value ones by 61 percent. Such differences in the offer are many times larger than design variations.
Target Audience & Segmentation (Priority 1)
Which recipients are contacted? RFM-based selection can increase CVR by 3.4x according to CMC 2025 — the single biggest lever.
Offer & Incentive (Priority 2)
Type, value, and mechanics of the coupon. CMC 2022 shows: percentage coupons +23% CVR, high-value +61% — massive differences.
Format & Response Boosters (Priority 3)
Letter vs. postcard, coupon card vs. QR code vs. scratch field. Physical inserts (coupon card: 6.1% CVR) outperform gamified elements.
Design & Envelope (Priority 4)
Layout, images, colors, envelope design. CMC 2025 shows: no measurable impact in B2C — content matters more than packaging.
Real-World Examples from the CMC Studies
The CMC Print Mailing Studies deliver the largest A/B tests in direct mail in the German-speaking region year after year. Since 2018, over 40 online retailers have sent more than one million print mailings to existing customers annually, with each edition focusing on a specific A/B test topic. The results set the direction for the entire industry.
The CMC Study 2023 tested four response boosters against each other: an enclosed coupon card, a printed coupon code, a QR code, and a scratch field. The result was clear-cut: the physical coupon card achieved a 6.1 percent CVR — the best value — 33 percent more than the scratch field (4.6 percent). The printed coupon code came in close behind at 5.5 percent. Overall, roughly 20 percent of mailing recipients visited the advertised online shop, with over 70 percent of visits occurring within the first two weeks. The 2020 study tested formats: the classic advertising letter outperformed maxi postcards and self-mailers by approximately 25 percent in CVR and achieved 42 percent higher ROAS than the self-mailer.
A particularly revealing result came from the CMC Study 2024 on coupon limitation: time-limited and quantity-limited coupons drove short-term responses massively upward (+38 percent for a 3-week deadline, +58 percent for quantity limitation) but lost out to the unlimited coupon in the long run. For short-term sales promotions, limitations are therefore effective, but for long-term existing customer reactivation they are counterproductive. This result would not have been detectable without systematic A/B testing — gut feeling would likely have favored the limited variant.
A/B Test Results from the CMC Studies (2020–2025)
Methodology — How an A/B Test Works in Direct Mail
A methodologically sound A/B test in direct mail follows a clear process. It begins with the hypothesis: "A 15% coupon with a 3-week deadline achieves a higher CVR than one without a deadline." Next, the recipient list is randomly divided into two equally sized groups — the control group receives the proven mailing, the test group receives the variant with exactly one change. All other factors (timing, target audience, remaining design) stay identical. Both variants are sent on the same day.
The critical question is the sample size: each test variant needs enough recipients to deliver statistically reliable results. The rule of thumb is: the sample should be large enough to generate at least 100 responses per variant. At an expected response rate of 2 percent, that means at least 5,000 recipients per variant — at 1 percent, at least 10,000. The target confidence level is typically 95 percent (p < 0.05), meaning the probability of a chance result is below 5 percent.
Particularly important — and different from digital tests — is the observation window: the CMC studies show that with print mailings, roughly half of all orders only come in from week five onward after dispatch. Anyone who evaluates their test after two weeks captures only a fraction of the effect and risks making wrong decisions. A minimum observation period of six to eight weeks after dispatch is recommended. Only then can the winning variant be identified and scaled up as the new "control" — the next test then tests a new element against this new benchmark.
A/B Testing: Direct Mail vs. Digital
Common Mistakes in A/B Testing
Even experienced marketers make methodological errors that can invalidate A/B tests. The most common mistake is changing multiple variables at once — for example, altering coupon value, headline, and design in a single variant. In that case, it is no longer possible to determine which change influenced the result. The "One Variable at a Time" (OVAT) principle is therefore the most important methodological rule.
An equally widespread mistake is a sample size that is too small. Testing with only 500 recipients per variant at a 2 percent response rate yields just 10 responses — far too few for statistically significant conclusions. The observed difference could be pure chance. Premature evaluation is another typical error: with print mailings, half of all orders come in only after weeks. Deciding after two weeks may lead to the wrong choice. Other common mistakes include unequal test groups (not randomly split), missing tracking (no individual codes), and adopting results without significance testing.
A/B Testing with AutoLetter
AutoLetter makes A/B testing in direct mail simple and accessible. Instead of coordinating elaborate test designs with traditional mail houses, businesses can configure their campaign variants online and have them sent to randomly split recipient groups — with integrated response tracking for clean performance measurement. This enables data-driven validation of hypotheses about offers, design, or personalization, without having to manually coordinate the testing process.
Especially for businesses that want to systematically optimize their direct mail, AutoLetter offers a decisive advantage: shorter cycles between test and rollout, transparent cost overview per variant, and measurable results as the foundation for the next campaign. Every test makes the following campaign better — fully in line with Claude Hopkins' principle: No campaign without a test.
Optimize your direct mail with data
With AutoLetter, you can easily test different campaign variants with measurable results — for higher response rates and optimal ROI.
Try it for freeFrequently Asked Questions About A/B Testing
5 Fragen beantwortet
The rule of thumb is: at least 100 responses per variant. At an expected response rate of 2 percent, you need at least 5,000 recipients per test variant; at 1 percent, at least 10,000. The target confidence level is typically 95 percent. Industry experts recommend 5,000–10,000 recipients per test cell as a benchmark.
Significantly longer than digital channels. The CMC studies show that roughly half of all orders come in only from week five onward after dispatch. A minimum observation period of 6–8 weeks is therefore recommended. Including preparation (design, printing, list creation) and evaluation, you should plan for 8–12 weeks in total.
The test variables with the greatest impact first: 1. Target audience/selection — has the strongest effect on results (CMC 2025: 3.4x CVR through RFM targeting). 2. Offer/coupon — type and value of the incentive (CMC 2022: +61% CVR for high-value coupons). 3. Format — letter vs. postcard (CMC 2020: +25% CVR for letters). Design details such as envelope design have no measurable impact in B2C according to CMC 2025.
In an A/B test, only one variable is changed (e.g., coupon value) while everything else remains identical — simple, clearly interpretable, and with moderate sample size requirements (5,000–10,000 per variant). In a multivariate test, multiple variables are tested simultaneously (e.g., coupon, headline, and design), which creates many combinations and requires very large sample sizes (5,000–10,000 per combination). For most direct mail campaigns, the A/B test is the more practical method.
The annual CMC Print Mailing Study by Deutsche Post is the largest benchmark study for direct mail in the DACH region. Key A/B test results: coupon card beats scratch field by 33% CVR (2023), unlimited coupons win long-term over limited ones (2024), letter outperforms postcard by 25% CVR (2020), percentage coupons beat fixed-amount coupons by 23% CVR (2022), and envelope design has no significant impact in B2C (2025).
Verwandte Begriffe
Response Rate
Key metric in direct marketing that measures the percentage of recipients who respond to a marketing campaign.
Conversion Rate
The conversion rate measures the proportion of recipients who complete a desired action. In print mailings, B2C campaigns achieve an average CVR of 4.1% — significantly higher than digital channels.
Wastage
The portion of an advertising campaign that reaches people outside the target audience—ineffectively deployed budget that can be reduced through precise targeting.
Personalization
Data-driven adaptation of advertising messages to individual recipients — from personalized salutations to fully individualized content using Variable Data Printing.
Target Audience Segmentation
The strategic process of dividing markets into homogeneous buyer groups — crucial in direct marketing for minimal waste circulation and maximum response rates.
Weiterführende Artikel
Direct Mail ROI Calculator + Analytics Guide 2025
67% of direct marketers don't know their true ROI. Learn the 3 ROI formulas, professional tracking setup, and continuous optimization for maximum profit.
Artikel lesenDirect Mail Design & Layout: How Visual Psychology Doubles Your Response Rate
The design of a direct mail letter decides within the first two seconds whether it gets read or tossed. While 82% of marketing managers optimize the copy, they neglect the visual layout. This guide shows how color psychology, typography and thoughtful design can measurably double your response rate on A4 letters.
Artikel lesenNew Customer Acquisition 2025: 8 Strategies with 11.3% Response (from 0.95€)
New customer acquisition with direct mail achieves 5.7-11.3% response rate – 12x higher than email. Discover 8 proven strategies with AutoLetter from 0.95€ per letter.
Artikel lesen