Cromulent
← Back to blog

How A/B Tests Become Message Learning

Justin T. Huang

Most email teams test with serious intentions. They compare subject lines, sender names, send times, offers, and copy treatments. They declare a winner, update the playbook, and move on to the next campaign.

This discipline is valuable because it creates friction against pure preference. Without tests, teams are left with anecdotes, personal taste, and whatever result someone remembers most clearly. The difficulty is that the dashboard often encourages a broader inference than the test can support. One version won, so the principle behind that version starts to feel like something the team has learned.

Sometimes that conclusion is reasonable. Often, it requires more care.

The first reason is that an email is a combination of many choices. A message varies by subject line, offer, tone, emotional frame, proof, specificity, urgency, product focus, send time, audience, and lifecycle context. A two-arm A/B test compares two combinations from that larger space. The result may tell us which message performed better for that audience, at that moment, on that metric. It does not automatically explain which choice within the message carried the effect.

That distinction matters because marketing teams usually want to learn something more general than "A beat B." They want to know whether discount framing works for this audience, whether proof reduces uncertainty, whether urgency helps or hurts, whether curiosity attracts valuable attention, and whether the same idea should be used again. Those are questions about message features, not only message variants.

The average customer is a convenience

When a dashboard says Variant B lifted conversion by 3.2 percent, the number summarizes many different customer responses. That summary is useful for reporting, but it can hide the structure of the effect.

Variant B might increase conversion among recent browsers, reduce conversion among long-term subscribers, and have little effect on new subscribers. The aggregate result can still look positive if one group is large enough or responsive enough. If the winner is then sent to everyone, the team has acted on an average that may describe no customer particularly well.

This is a familiar problem in lifecycle marketing because different customer histories create different reasons to care. A repeat buyer may already trust the brand, which makes early access meaningful. A new subscriber may need reassurance about fit, quality, returns, or whether the product solves the problem that brought them in. A lapsed buyer may need to be reminded why the category matters before a promotion has much force.

The winner still matters. It gives the team an answer to the question it asked. The variation around the winner often teaches something more durable: where the message worked, where it did not, and which customer contexts changed the result.

Copy is a bundle of features

The same issue appears inside the message itself. A variant is usually a bundle of choices, even when the test is described as a simple comparison.

Consider two subject lines.

A: "Your exclusive early access starts now."

B: "20% off our bestselling jacket this weekend."

If B wins, the team may conclude that the discount worked. That may be right, but the comparison changed more than the discount. B is more product-specific. It includes a popularity cue. It creates a time window. It may attract readers who were already interested in the jacket. The dashboard can report which bundle performed better, but interpretation depends on how the bundle is described.

This is where structure helps. Before a result can travel from one campaign to the next, the team needs some vocabulary for what the message contains: discount, exclusivity, product specificity, social proof, urgency, direct address, authority, reassurance, or other mechanisms relevant to the campaign. The labels will be imperfect, and they should not turn copywriting into taxonomy for its own sake. They simply need to be consistent enough that future campaigns can reuse the learning.

Once variants are described this way, isolated tests begin to form a more useful record. Proof may matter more for first-time buyers than for repeat buyers. Urgency may help replenishment reminders while adding little to educational campaigns. Discounts may raise clicks without improving margin enough to justify the cost. Curiosity may raise opens while selecting readers with weaker purchase intent.

The practical value of testing grows when the team can connect performance back to the message features that plausibly produced it.

Sequential learning

Once variants have structure, each campaign can inform the next one more directly. A result no longer has to live only as "A beat B." It can update a more useful belief: proof seems to reduce uncertainty for this segment, utility may work nearly as well as discount for this product, urgency may be attracting attention without changing purchase intent.

This way of working is Bayesian in spirit, though the terminology is less important than the habit. A campaign begins with beliefs about customers and messages. Customers respond. The beliefs are revised, sometimes slightly and sometimes substantially. The next campaign should begin from the revised state rather than from the same starting assumptions.

Different outcomes can also imply different lessons. A message that raises opens may not raise purchases. A subject line that attracts attention may not attract intent. A discount frame may increase clicks while doing less for margin than a clearer product claim. Treating these outcomes separately keeps the team from collapsing every form of response into one vague idea of performance.

In many teams, the useful part of a campaign result survives as screenshots, Slack comments, or recap slides. Six weeks later, the next campaign begins from a mixture of memory and intuition. A sequential testing program tries to keep the important pieces together: the audience, the message features, the outcome, and the interpretation.

Suppose proof performs better than price among first-time buyers. The next question might be whether that pattern holds across product categories. Suppose utility-oriented copy produces similar revenue to discount-oriented copy while reducing unsubscribes. The next question might be whether utility can protect margin in other segments. Suppose urgency lifts clicks but not revenue. The next question might be whether urgency is creating attention without increasing purchase intent.

Each result narrows the next question. Campaign history becomes more useful when it records what the brand has learned about messages, customers, and conditions.

What carries forward

A two-arm test is most useful when its claim is kept modest. Given this audience, offer, timing, creative bundle, and metric, one constructed message outperformed another. That sentence is less dramatic than a victory slide, but it is closer to what the test actually showed.

The value comes from accumulating many such claims with their context intact. Over time, the brand learns where clarity is more valuable than curiosity, where proof matters more than price, where urgency deserves restraint, and where the average hides important differences across customers.

The aim is modest and valuable: after each send, the next campaign should begin with fewer unsupported assumptions. That requires more than a record of which variant won. It requires a way to connect message features, audience context, outcomes, and interpretation, so that each campaign changes what the next one is likely to try.


Cromulent helps marketing teams turn campaign execution into a structured learning process. The system generates copy from explicit persuasion hypotheses, tests those hypotheses in live campaigns, and updates its models as performance data arrives. Each send becomes a source of evidence about which messages work, for which audiences, and under which conditions.