Incrementality or Illusion: Getting Past Last‑Click in 90 Days
Why “last‑click” keeps lying to you
Trending Category
The money looks good on a dashboard built for last‑click. Search appears to print sales; social seems to freeload. Then a CFO asks a simple question: What would have happened without the spend? That’s where the illusion cracks. In one widely cited field experiment, eBay found branded search produced virtually no incremental revenue, with organic traffic replacing almost all paid clicks.¹ Yet Google’s own pause studies reported that when advertisers turn off search ads, roughly 89% of the lost paid clicks aren’t recovered by organic.² Both are true—because context is doing the heavy lifting. The only way to resolve the paradox is to measure causally, not correlationally.
This is the 90‑day, mid‑market playbook to do exactly that—using geo‑holdouts, PSA/“placebo” controls and ghost‑ad logic, and a triangulated MMM that is calibrated to real experiments. It’s practical, budget‑aware, and defensible in an executive meeting.
Why “last‑click” keeps lying to you
Attribution stitched from click paths is correlation masquerading as causation. As eBay’s experiment showed, high‑intent users type a brand and will arrive with or without the ad; last‑click credits the ad anyway.¹ Conversely, Google’s pause meta‑analysis spans many advertisers and queries, finding that in most cases not advertising does reduce total clicks—often dramatically.² The lesson isn’t that one side is wrong. It’s that lift varies by brand equity, query type, and audience—and you only see it with experiments that manipulate exposure.
Two innovations made this practical at mid‑market scale:
- Ghost‑ad frameworks match “would‑have‑seen” users to exposed users so you can compute lift without buying placebo impressions for every control user.³
- Geo experiments randomize advertising at the level of cities/DMAs, then estimate lift with pre‑test baselines and synthetic controls; they’re conceptually simple and privacy‑durable.⁴ ⁵
The simplest defensible design for a mid‑market budget (90 days)
Audience: brands spending roughly five-to-low‑six figures per month across paid search, paid social, and a video channel (YouTube/CTV).
Goal: credible incremental ROAS (iROAS) by channel with 90% confidence intervals, plus clear creative implications.

Phase 1 (Weeks 0‑2): Instrument & pre‑test
- Pick one north‑star outcome (e.g., qualified lead, first purchase) and freeze conversion definitions.
- Pull 12–16 weeks of historical, weekly data by geo (sales, traffic, store density, seasonality).
- Establish pre‑test baselines for geos—Google’s own guidance suggests 4–8 weeks of pre‑test when running geo experiments.⁶
Phase 2 (Weeks 3‑6): Run a geo‑holdout on your highest‑spend channel
- Randomize 8–12 DMAs into test/control and hold spend flat elsewhere.
- Use GeoLift (synthetic control) to select balanced markets, run power analysis, and compute lift with jackknife+ or conformal intervals.⁷ ⁸
- Expect wider CIs with few geos; underpowered tests will swing from +2% to +50% “lift” just on noise.⁹
- If your stack supports it, Google’s Conversion Lift (geography) can complement the open‑source workflow, though it tends to require higher budgets than user‑level lift.¹⁰
Deliverables: incremental conversions, iROAS, and 90% CI for Channel A, plus a sanity check that test and control reconverge after the test window.¹¹
Phase 3 (Weeks 5‑8): Layer a people‑level lift where feasible
- On Meta or YouTube, run platform Conversion Lift (sales outcome) or Brand Lift (creative outcome). Lift tests typically report at 90% confidence and handle user‑level randomization.¹² ¹³ ¹⁴
- If you must keep auction dynamics identical across cells, a PSA/placebo control can work (showing controls a charity PSA). But be careful: platform optimizers often treat PSAs differently, biasing reach and wrecking comparability; ghost‑ad style controls avoid that pitfall.³ ¹⁵ ¹⁶ ¹⁷
Deliverables: channel‑specific incremental conversions and cost‑per‑incremental with CIs; for video, brand‑metric direction by creative.
Phase 4 (Weeks 6‑10): Triangulate with MMM—but calibrate it
- Fit a lightweight MMM (Meta’s Robyn or Google’s Meridian) on weekly aggregates with adstock and saturation to get cross‑channel response curves.¹⁸ ¹⁹ ²⁰ ²¹
- Calibrate MMM using your experimental ground truth (Robyn includes an MAPE.LIFT objective for this; Meridian explicitly supports integration of lift results).²² ²¹
- Output mROAS by spend level to plan the next quarter.
Deliverables: calibrated MMM with credible intervals and a budget reallocation recommendation consistent with your lift results.
What results to expect (and how to read them)
Lift by channel. Expect a spectrum. For high‑intent brand search, large incumbents sometimes see near‑zero short‑term lift (users would have arrived anyway), while other advertisers show strong incrementality—Google’s meta‑analysis averages ~89% incremental clicks when search is paused.¹ ² Read the local result, not the average.
Confidence intervals. Most platforms report 90% CIs for lift; your internal standard should be 90–95% depending on risk tolerance. A CI like “+12% ± 6%” is actionable; “+26% ± 24%” is not.¹² ¹⁴ ⁹
Creative effects. Don’t ignore creative. Controlled Brand Lift studies routinely show winners and losers among assets; picking the right asset can meaningfully change downstream conversion lift.¹³
Geo‑holdouts, PSA tests, MMM: when to use what
Technique | Best For | Strengths | Watch‑outs |
Geo‑holdout (open‑source GeoLift or Google geo lift) | Search, retail, multi‑region DTC | Privacy‑durable; measures real sales; works cross‑platform | Needs enough geos and pre‑test; power can be low with small clusters.⁷ ⁸ ⁶ |
People‑level lift (Meta/YouTube) | Social/video | True RCT; clean randomization; platform‑managed samples | Requires platform budgets; focus on 90%+ confidence.¹² ¹³ |
PSA/placebo control | CTV, walled gardens without lift | Keeps ad load constant in both cells | Can bias audience mix; extra media cost; use cautiously or prefer ghost‑ad style controls.¹⁵ ¹⁶ |
Ghost‑ad logic | Display, retargeting at scale | Matches “would‑have‑seen” users without buying PSAs | Requires platform/partner support and tight identity logic.³ |
MMM (Robyn/Meridian) | Portfolio planning | Holistic view, adstock and saturation; weekly cadence | Correlational without calibration—anchor to experiments.¹⁸ ¹⁹ ²² ²¹ |
A CFO‑ready readout (what to put on one slide)
- iROAS by channel with 90% CI (from lift).
- Cost per incremental outcome vs. last‑click CPA.
- Creative winner (if video) from Brand Lift.
- Calibrated MMM reallocation: “Move 15–25% from Channel B to Channel A at current spend levels” with expected range.
- Next test queued (e.g., holdout on retargeting, PSA‑controlled CTV).
Common failure modes (and how to avoid them)
- Underpowered tests. Too few geos or too short a test produce wide CIs; run power analysis first.⁸ ⁹
- Biased controls. PSA ads can distort who you reach; Google has shown how optimizers treat PSAs differently. Prefer platform lift or ghost‑ad logic.¹⁵ ¹⁶
- Cookie contamination. Running overlapping platform tests can pollute each other’s control cells; stagger or split by geo.¹⁸
- Un‑calibrated MMM. Treat MMM as triangulation, not truth—calibrate to experiments or don’t use it for budget moves.²² ²¹
Bottom line: Last‑click is a story; incrementality is evidence. In 90 days, you can replace the former with the latter—and move the budget with conviction.
Endnotes
- Thomas Blake, Chris Nosko, Steven Tadelis, “Consumer Heterogeneity and Paid Search Effectiveness: A Large‑Scale Field Experiment,” Econometrica / NBER working paper (eBay brand vs. non‑brand effectiveness). (Haas School of Business, NBER)
- David X. Chan et al., “Incremental Clicks Impact of Search Advertising” (Search Ads Pause meta‑analysis; ~89% incremental clicks on average). (Google Business, Google Research)
- Garrett Johnson, Randall Lewis, Elmar Nubbemeyer, “Ghost Ads: A Revolution in Measuring Ad Effectiveness,” Journal of Marketing Research (ghost‑ad methodology for causal lift at auction scale). (SAGE Journals)
- Jon Vaver, Jim Koehler, “Measuring Ad Effectiveness Using Geo Experiments,” Google Research (geo randomization across markets). (Google Research, Google Research)
- GeoLift documentation (synthetic control methods; open‑source geo experimentation workflow). (Facebook Incubator)
- Unofficial Google Data Science blog, “Estimating causal effects using geo experiments” (structure with 4–8 weeks pre‑test). (Unofficial Google Data Science)
- GeoLift Walkthrough (market selection, power calculations, execution). (Facebook Incubator)
- GeoLift methodology for confidence intervals (conformal and jackknife+). (Facebook Incubator)
- Polar Analytics, “Why Most Geo Tests Fail” (interpretation of wide CIs; power constraints with small geo counts). (Polar Analytics)
- Google Ads Help, “Conversion Lift based on geography” (budget considerations; aggregated by non‑overlapping regions). (Google Help)
- Expedia Group Tech, “Measuring marketing success: incrementality & geo‑testing” (post‑test reconvergence and CI reporting). (Medium)
- Meta for Developers, “Conversion Lift Measurement” and Meta Business Help, “About confidence in your tests” (lift methodology and 90%+ confidence guidance). (Facebook Developers, Facebook)
- Google Support (DV360) “Understand your Brand Lift measurement data” and YouTube Brand Lift case guidance (creative study design and thresholds). (Google Help)
- Google Ads Help, “Lift measurement statuses and metrics” (reporting with 90% confidence intervals). (Google Help)
- Think with Google, “A Revolution in Measuring Ad Effectiveness” (how PSA controls can bias auctions; why ghost‑ad logic is superior). (Google Business)
- Remerge, “Incrementality Tests 101: ITT, PSA and Ghost Ads” (PSA pros/cons; cost and bias trade‑offs). (Remerge)
- Viant, “Lift & Incrementality in Programmatic Advertising” and “Why Your Control Group Could Be Ruining Incrementality” (PSA/ghost‑ad overview; control comparability). (Viant Technology LLC)
- Meta’s Robyn MMM (open source; adstock, saturation; documentation). (Facebook Experimental)
- Google’s Meridian MMM (open source; launched 2025; integration path). (Google for Developers, Google Business)
- Google’s LightweightMMM (Bayesian MMM foundations). (GitHub, LightweightMMM Documentation)
- eMarketer coverage of Meridian’s aims (triangulation with experiments and attribution). (EMARKETER)
- Robyn calibration to ground truth (MAPE.LIFT objective to minimize gap to experimental lift). (Facebook Experimental)
The Blog
Step-by-step instructions on defining and documenting the tone of voice for your startup Your Voice Is Your Brand—Before Your Product…
Why Your Name Matters More Than You Think Naming your startup isn’t just a creative decision—it’s a growth lever. The…
A guide to building logos, color palettes, and fonts for a DIY brand kit without expensive designers. Why a DIY…