The testing velocity playbook. 6 tests. Per quarter.
The median CRO team ships 2 tests per quarter. The top 10% ship 6–8. The difference isn't brains or budget — it's operational. Here's what the fast teams do that the rest don't.
- ▸Testing velocity is a function of five bottlenecks: ideas, prioritization, spec, build, ship. Almost every team is stuck on one.
- ▸The highest-leverage fix is almost never 'more testing tools' — it's a visual editor that unblocks marketing from engineering sprints.
- ▸Running a weekly test council with forced-rank prioritization eliminates the 'what do we test next?' meeting permanently.
- ▸Auto-promotion at 95% CI removes the biggest source of lost velocity: the decision meeting after a winner lands.
Find where your team is stuck.
Every stuck testing program is stuck on exactly one of five bottlenecks. Identify yours, fix it first, and velocity doubles. Trying to fix all five at once is how teams stay at 2 tests per quarter for 18 months.
- ▸Idea generation — too few hypotheses on the backlog.
- ▸Prioritization — hypotheses exist but nobody agrees which to run.
- ▸Spec — prioritized tests stall because the brief is ambiguous.
- ▸Build — specs are clear but engineering doesn't have capacity.
- ▸Ship — the test is built but deploy requires a release train.
If your weekly standup spends more than 10 minutes debating which test to run next, your bottleneck is prioritization — not ideas.
Where the best hypotheses come from.
The best testing ideas don't come from brainstorms. They come from three sources, in this priority order: user research, competitor intel, and funnel analytics. Teams that run all three see 3x more qualified hypotheses on their backlog than teams that rely on quarterly brainstorms.
- ▸User research: 5 interviews per quarter, transcribed, tagged by friction point.
- ▸Competitor intelligence: every positioning/pricing move your top 5 competitors make = a test hypothesis.
- ▸Funnel analytics: any drop-off bigger than 15% between two steps is a test candidate.
- ▸Avoid brainstorms until the above three are running — you'll generate ideas faster and with higher hit rates.
The weekly test council.
Kill the 'what do we test next?' debate by making it a scheduled event with forced-rank output. 30 minutes, once a week, with the CRO manager, one engineer, and one marketer. Every hypothesis gets a score; the top 3 go into the sprint. No exceptions.
- ▸Use ICE (impact, confidence, ease) or PIE scoring — any consistent framework.
- ▸Force-rank the top 10 hypotheses every Monday. No ties, no 'we'll decide later'.
- ▸Anything outside the top 10 goes to the parking lot — reviewed quarterly.
- ▸The CRO manager holds the final tie-break vote. Committees kill velocity.
Decouple marketing from the release train.
The single biggest velocity killer we see: every test requires an engineering ticket, a sprint slot, and a deploy. Tools with a visual editor (or a WordPress / component-level plugin) let marketing ship tests without engineering bandwidth. Teams that make this transition typically go from 2 tests/quarter to 6+ within one quarter.
- ▸Visual editor for copy, CTA, and layout tests — marketer-owned.
- ▸Engineering-owned tests are reserved for logic and pricing changes.
- ▸Auto-promotion at 95% CI means winners ship the same day — no decision meeting.
- ▸Rollback in one click — removes the 'what if we break prod?' objection.
Measure what matters weekly.
What gets measured moves. Four metrics, reviewed every Monday, make velocity non-negotiable. Teams that track these four hit the top decile; teams that don't, don't.
- ▸Tests shipped per week (target: 0.5+).
- ▸Average time-from-hypothesis-to-ship (target: <14 days).
- ▸Hit rate on winning tests (target: 20–30% — higher means you're not being ambitious).
- ▸Revenue impact per test shipped (target: $10K+ in incremental annualized revenue).
Get the velocity scorecard template.
Google Doc, Notion, and PDF versions. Shared with you in one email.
Download →Related for your role
ALL RESOURCES →Stack Consolidation ROI Calculator
Enter what you pay Optimizely, Crayon, Hotjar, and Ahrefs today. See what Optimize Pilot would cost instead — and how many headcount the delta covers.
Experiment Brief Template
The one-page brief every test should start from: hypothesis, variants, sample size, success metric, guardrails. Notion + Google Doc versions inside.
Statistical Significance, Actually Explained
P-values without the jargon. When to use Frequentist vs Bayesian. When to stop a test early without lying to yourself. A practical primer for CRO teams.
Hit 6 tests a quarter. Inside the trial.
Most teams that run the Optimize Pilot trial ship 3–4 tests in the 14 days. Navigator AI removes the prioritization meeting; the visual editor removes the engineering dependency.