Campaign Testing Phases: Sprint, Test, Scale (The Complete Framework)
Mitchell Keller
Founder & CEO, LeadGrow · Managed 3,626+ cold email campaigns. 6.74% average reply rate. Booked 2,230+ meetings in 2025.
TL;DR
- 12 variants across 3 angles. 1,000 contacts max. Weekly cycles. Find the winning angle, not the winning words.
- 3 to 6 variants of the winner. Accept slight performance drops for longevity. Harden the angle into a durable sequence.
- 80% to proven messaging. 20% to testing adjacent angles. This is where meetings compound.
- Start refilling contacts at 30% sent through. Complete by 50%. Never run dry.
- **12% positive reply rate is the floor** before scaling. Below that, you're scaling mediocrity.
By Mitchell Keller, Founder & CEO, LeadGrow. Managed 3,626+ cold email campaigns. 6.74% average reply rate. 2,230+ meetings booked in 2025.
Most Teams Scale Before They've Found Anything Worth Scaling
The typical cold email campaign testing process looks like this: write an email, send it to 5,000 people, check the numbers after 2 weeks, tweak the subject line, send again. When it doesn't work, blame the copy. Or the list. Or deliverability. Or cold email as a channel.
The problem is not that cold email doesn't work. It's that most teams skip the testing phase entirely and jump straight to scale. They're sending 500 emails a day before they know which angle resonates, which pain point triggers replies, or which positioning gets meetings.
That's like running a restaurant and serving every dish on the menu before figuring out which ones customers actually order.
Across 3,626+ campaigns, we've developed a 3-phase testing framework that systematically finds the winning angle, hardens it against fatigue, and then scales it for maximum output. This post teaches the exact framework.
Phase 1: Sprint (Month 1)
The Sprint phase has one job: find the angle that resonates with your market. Not the exact words. Not the perfect subject line. The angle.
An angle is the frame through which you present your offer. Same product, different lens.
Example for a sales intelligence tool:
- Problem angle: "Your reps are wasting 5 hours a week researching accounts manually"
- Peer angle: "Companies like [competitor] are using AI to prep for calls in 2 minutes"
- Efficiency angle: "Cut account research from hours to minutes"
Each angle speaks to the same buyer with the same product. But they hit different motivations: pain avoidance, competitive pressure, and time savings. One of these will resonate 2 to 5x more than the others. The Sprint phase finds which one.
The 12-Variant Sprint Structure
We launch 12 email variants simultaneously. That's 3 positioning angles with 4 variants per angle (different subject lines, opening hooks, or CTAs within each angle).
Why 12? Because 3 angles with 1 variant each doesn't give you enough data to separate angle performance from copy performance. 4 variants per angle lets you see whether the angle itself works, even if individual copy variations underperform.
If 3 out of 4 variants in the "problem angle" outperform all 4 variants in the "peer angle," you've found your winning angle. That conclusion holds even if the fourth problem-angle variant was weak.
Sprint Constraints
- 1,000 contacts maximum. You're testing, not sending at scale. Burning through a huge list before finding the right angle wastes your best prospects on unproven messaging.
- Weekly cycles. Each week, review the data, kill the bottom performers, and launch new variants in the winning angle's direction. 4 weekly cycles in Month 1 gives you 4 iterations.
- Equal distribution. Each variant gets the same number of sends. If one variant gets 300 sends and another gets 50, the data is useless. Equal distribution makes comparisons valid.
- 200 to 300 sends per variant minimum. Anything less and random variation makes the data unreliable. 250 sends per variant across 4 variants per angle = 1,000 contacts total. That's why 1,000 is the target.
What to Measure (And What to Ignore)
During Sprint, the only metric that matters is positive reply rate. Not open rate. Not click rate. Not total reply rate.
Open rates tell you about subject lines, not angles. Click rates tell you about CTAs, not positioning. Total reply rates include "not interested" and "remove me" responses, which inflate the number without indicating market fit.
Positive reply rate counts only responses that express interest, ask a question, or engage with your offer. That's the signal. Everything else is noise during Sprint.
The Sprint Hard Decision
At the end of Month 1, you make a hard decision. There are three possible outcomes:
Outcome 1: Clear winner (one angle at 12%+ positive replies). Move to Phase 2. You've found the angle.
Outcome 2: Promising but not clear (one angle at 8 to 12%). Run one more Sprint week with refined variants of the top angle. If it climbs above 12%, move to Phase 2. If it doesn't, expand the test to new angles.
Outcome 3: Nothing works (all angles below 8%). This means one of three things: your list quality is wrong, your offer doesn't match the market, or your ICP definition needs work. Do not move to Phase 2. Go back to foundations. We've killed campaigns at this stage and restarted with different ICP definitions. It's painful, but scaling a campaign with sub-8% positive replies just burns through your market faster with bad messaging.
Real Sprint Example
Client: B2B SaaS selling to mid-market finance teams.
Sprint setup: 3 angles (compliance pain, efficiency gain, competitive pressure). 4 variants each. 1,000 contacts.
Week 1 results:
- Compliance pain: 4.2%, 5.1%, 3.8%, 4.9% positive reply rates
- Efficiency gain: 8.3%, 9.1%, 7.6%, 8.8% positive reply rates
- Competitive pressure: 2.1%, 1.9%, 3.2%, 2.4% positive reply rates
Efficiency angle won clearly. We killed compliance and competitive angles. Weeks 2 through 4 tested 8 new variants within the efficiency angle, refining the hook and CTA. Best variant hit 13.2% by end of Month 1. Clear winner. Moved to Phase 2.
Phase 2: Winner-Testing (Month 2+)
Phase 2 takes the winning angle from Sprint and hardens it. You've found what resonates. Now you need to make it durable.
A single email variant will fatigue over time. The same people in your market will see the same messaging. Industry peers will compare notes. What worked at 13% in Month 1 might drop to 8% by Month 3 if you don't evolve it.
The 3 to 6 Variant Expansion
Take your winning angle and create 3 to 6 new variants that express the same positioning in different ways:
- Different opening hooks within the same angle
- Different proof points (different client examples, different metrics)
- Different CTAs (meeting vs resource vs question)
- Different sequence structures (2-step vs 4-step vs 6-step)
- Different personalization approaches (company-level vs role-level vs situation-level diagnosis)
The angle stays the same. The execution varies. This gives you multiple "weapons" within the winning frame so you can rotate them and fight fatigue.
Accept the Performance Dip
When you expand from your top Sprint variant to 3 to 6 new variants, average performance will drop slightly. Your Sprint winner was the absolute best performer from a month of testing. The new variants haven't been optimized yet.
This is expected. If your Sprint winner hit 13%, your Phase 2 variants might average 9 to 11%. That's fine. You're trading peak performance on one variant for sustainable performance across multiple variants. The portfolio approach outlasts the single-variant approach every time.
Phase 2 Benchmarks
- Target: 3+ variants performing above 8% positive reply rate
- Timeline: 4 to 6 weeks to stabilize
- Contact volume: 2,000 to 5,000 contacts across all variants
- Decision point: When you have 3+ stable variants above 8%, move to Scale
Phase 2 Kill Criteria
Any variant below 6% positive replies after 300 sends gets killed. Don't give it more time hoping it improves. Replace it with a new variant within the winning angle. You should be rotating through new variants continuously, keeping the ones that work and replacing the ones that don't.
Phase 3: Scale
Scale is where campaigns start printing meetings. You've found the angle (Sprint). You've built durable variants (Winner-Testing). Now you increase volume confidently because you know what works.
The 80/20 Split
80% of your sending volume goes to proven variants. 20% goes to testing adjacent angles. This ratio protects your meeting flow while keeping the campaign evolving.
The 20% test budget matters. Markets change. Competitors enter. Pain points shift. What worked 3 months ago might not work 6 months from now. The testing budget is your insurance policy against market shifts.
Scaling Tiers
How aggressively you scale depends on your positive reply rate:
| Positive Reply Rate | Scaling Approach | Volume Increase |
|---|---|---|
| 20%+ | Aggressive | Double volume weekly until capped by TAM or infrastructure |
| 12 to 20% | Significant | 50% volume increase every 2 weeks |
| 8 to 12% | Steady | 25% volume increase every 2 weeks |
| 6 to 8% | Cautious | Hold current volume, optimize before scaling |
| Below 6% | New Sprint | Stop scaling. Return to Phase 1 with new angles |
12% positive reply rate is our floor before scaling. Below that, you're scaling mediocrity. The math doesn't compound favorably because you're burning through your market faster than you're converting it. Fix the conversion first, then add volume.
The 30/50 Refill Rule
This is the operational detail that most teams miss. When 30% of your contact list has been sent to, start refilling the list with fresh contacts. Complete the refill by the time 50% of the original list has been sent to.
Why? Two reasons:
- You never want to run out of contacts mid-campaign. A campaign that runs dry loses momentum, sending reputation, and timing consistency. The 30/50 rule ensures continuous fuel.
- Fresh contacts maintain performance. The people on your original list who haven't been contacted yet are the ones least likely to convert (you've already reached the highest-fit ones). Refreshing with new, freshly sourced contacts keeps quality high.
In practice, this means your list building is not a one-time event. It's a continuous process that runs parallel to your campaigns. We typically rebuild lists every 4 to 6 weeks, using the same Clay workflows with updated filters and fresh data.
When to Kill a Campaign
Not every campaign survives the framework. Knowing when to kill is as important as knowing how to scale.
Kill Signal 1: Sprint Fails Twice
If you've run two full Sprint cycles (8 weeks of testing) with different angles and no variant has cracked 8% positive replies, the problem is upstream. Your ICP is wrong, your offer doesn't resonate, or your product isn't ready for this market. Stop burning sends and fix foundations.
Kill Signal 2: Declining Performance Despite Fresh Variants
If Phase 2 variants that initially performed at 10%+ are all declining below 6% despite continuous rotation, the market is fatiguing. Either your TAM is too small (you've saturated the reachable audience) or the competitive landscape has shifted.
Kill Signal 3: Meetings Don't Convert to Revenue
This is the sneaky one. A campaign might book 15 meetings per month with a 12% reply rate, but if none of those meetings turn into revenue, the campaign is a vanity metric generator. Reply rates and meeting counts only matter if they connect to pipeline and closed deals.
We track meeting-to-opportunity conversion for every campaign. If the conversion rate is below 20% (meaning fewer than 1 in 5 meetings becomes a real opportunity), something is wrong with the qualification layer. The emails are reaching people who respond but aren't real buyers.
Real Testing Pivot Examples
Pivot 1: From Product Features to Competitive Displacement
Client selling project management software. Sprint tested: feature comparison, productivity gains, and team alignment angles. All underperformed (3 to 5% positive replies).
We noticed in the few positive replies that people kept mentioning dissatisfaction with their current tool. Pivoted to a "switching without disruption" angle. Next Sprint hit 11.4%. The market didn't need to be sold on project management. They needed permission to switch.
Pivot 2: From Decision Maker to Influencer
Client selling to enterprise IT departments. Targeting CIOs. Sprint Phase showed 2.1% positive replies. CIOs were too insulated from the day-to-day problem the product solved.
Pivoted to targeting IT Managers and Directors (the people who actually feel the pain daily). Same angle, different title targeting. Positive reply rate jumped to 9.8%. Sometimes the angle is right but the audience is wrong.
Pivot 3: From Cold to Event-Triggered
Client in the HR tech space. Standard cold outreach across all angles peaked at 4.7% positive replies. Market was crowded and fatigued.
Pivoted to event-triggered outreach. Targeted companies that posted new DEIB related job openings (indicating focus on the exact problem the product solved). Same positioning angle, but sent only to companies with the hiring signal. Reply rate: 16.2%. The signal made the message relevant. This is situation mining in action.
Operational Details That Matter
Testing Velocity
Speed kills mediocrity. The faster you test, the faster you find what works. We aim for 4 to 8 new variant tests per week during Sprint. That's aggressive, but it compresses the learning timeline from months to weeks.
Lower testing velocity means longer time to answers, which means more budget spent on unproven messaging. If you can only manage 2 tests per week, your Sprint phase takes 2 months instead of 1. Plan accordingly.
Statistical Significance
Don't overcomplicate this. With 200 to 300 sends per variant, you won't reach traditional statistical significance. That's fine. You're looking for directional signals, not academic proof. If one angle is at 12% and another is at 4% after 250 sends each, you don't need a p-value to know which one to pursue.
The framework compensates for statistical imprecision through iteration speed. Test fast. Cut losers fast. Double down on winners fast. The rapid feedback loop self-corrects for random variation.
Documentation
Track every test. Angle, variant copy, subject line, audience segment, send volume, reply rate, positive reply rate, meetings booked. Store this in a spreadsheet or project management tool that's searchable.
Six months from now, when a new campaign in a similar vertical needs angle ideas, your testing history is a goldmine. We've built an archive of 3,626+ campaigns' testing data. It's the reason new campaigns reach Phase 2 faster. We're not starting from zero. We're starting from patterns.
The Bottom Line
Campaign testing is not optional. It's the entire strategy. The Sprint, Test, Scale framework is how you systematically find what works instead of guessing and hoping.
Sprint finds the angle. Winner-Testing hardens it. Scale prints meetings. Skip any phase and the whole thing falls apart.
The teams that book 2 to 5 meetings per week from cold email are not running magic copy. They're running a disciplined testing process that compounds over time. The first campaign takes a full month to find the angle. The tenth campaign in the same vertical finds it in a week because you've built pattern recognition from the previous nine.
That compounding is the real advantage. It's not about any single campaign. It's about the system. If you're building this system from scratch, our cold email writing guide covers the copywriting fundamentals that feed into every variant you'll test.
Key Statistic: Campaigns that complete all 3 phases (Sprint, Test, Scale) average 2.8x more meetings per month than campaigns that skip the Sprint phase and go directly to volume, based on LeadGrow's data across 3,626+ campaigns.
>
Source: LeadGrow internal campaign data, 2025
Frequently Asked Questions
Related Articles
Want us to run this playbook for you?
Book a strategy call and we'll show you how these frameworks apply to your business.
Book Strategy Call