Back to Guides
Attribution18 min read8 chapters

The Complete Guide to Marketing Attribution for Ecommerce

Why platform-reported ROAS is wrong, how holdout testing works, and how to find true incremental value per channel. The framework behind every attribution decision Parker runs on Cresva.

Parker's Domain

Chapter 1: The Attribution Crisis

Marketing attribution is broken. Not slightly off - fundamentally, structurally broken. The numbers your platforms report are systematically inflated, and the entire industry allocates billions of dollars based on data that overclaims by 23% on average.

23%

Average Overclaim

Cross-platform

28%

Meta Inflation

Avg across verticals

18%

Google Inflation

Including branded

35%

TikTok Inflation

Highest variance

Here's the core problem: every ad platform is both the seller of advertising AND the measurer of advertising effectiveness. Meta tells you Meta works great. Google tells you Google works great. TikTok tells you TikTok works great. And because they all use different attribution methodologies - view-through windows, click windows, self-attributed conversions - you can add up all the platform-reported conversions and get a number far higher than your actual total conversions.

We call this the “sum problem.” If Meta claims 1,000 conversions, Google claims 800, and TikTok claims 400, that's 2,200 total. But your Shopify shows 1,500 actual orders. Someone is wrong. In reality, everyone is wrong - they're all overcounting, just by different amounts.

The platforms that sell you ads are the same ones measuring whether those ads work. This fundamental conflict of interest means every ROAS number you see is inflated. The question isn't whether it's wrong - it's how wrong.

Chapter 2: Why Platforms Lie

“Lie” is strong. The platforms aren't deliberately fabricating numbers. They're using attribution methodologies that systematically favor themselves. There are four primary mechanisms:

View-Through Attribution

Meta counts a conversion if someone saw your ad and purchased within the attribution window - even if they never clicked, never engaged, and would have purchased anyway. If someone scrolls past your ad at 2am and buys from a Google search at noon, Meta claims that conversion.

Impact: High - accounts for 40-60% of Meta's overclaim

Multi-Platform Double Counting

A customer sees a Meta ad, clicks a Google ad, and buys. Both platforms claim the full conversion. Neither reports 0.5 conversions. The same sale is counted twice.

Impact: Medium - affects 15-25% of conversions

Organic Cannibalization

Your most loyal customers were going to buy anyway. But they happened to see an ad or click a branded search result on the way to your site. The platform claims that sale as ad-driven.

Impact: High - especially for branded search campaigns

Algorithmic Attribution Windows

Platforms use different attribution windows (1-day, 7-day, 28-day) and default to the most generous. Longer windows capture more coincidental correlations, not causal relationships.

Impact: Medium - inflates by 10-20% depending on window

The iOS 14.5 Factor

Apple's ATT framework made tracking harder, but it didn't fix attribution - it made platforms more creative about claiming conversions. Modeled conversions, probabilistic matching, and broadened attribution windows mean the overclaim problem actually got worse post-iOS14, not better. Platforms now “estimate” conversions they can't directly track, adding another layer of inflation.

Chapter 3: Attribution Models Compared

Before diving into solutions, you need to understand the landscape. There are four main attribution approaches, each with distinct tradeoffs. The industry is shifting from simpler models toward incrementality-based approaches.

Attribution Model Comparison

How it works

100% credit to the last touchpoint before conversion.

Strengths

Simple, easy to implement, no ambiguity

Weaknesses

Ignores discovery channels, heavily biases toward branded search and retargeting

Our Verdict

Massively undervalues awareness and consideration. Will lead you to over-invest in bottom-funnel.

The ideal approach combines methods: use MMM for strategic quarterly allocation, incrementality testing for validating channel effectiveness, and corrected MTA for daily optimization. No single model is sufficient on its own.

What Parker Does

Parker uses a hybrid approach - running continuous incrementality calibration against platform-reported data, applying correction factors per channel, and feeding corrected numbers to Felix's forecasting models. The result: attribution numbers you can actually trust for budget decisions.

Chapter 4: Overclaim by Platform

Not all platforms overclaim equally. Based on holdout testing across 100+ ecommerce brands on Cresva, here are the typical inflation rates:

Platform Overclaim Calculator

See how much your platform is likely inflating ROAS.

Reported

4.2x

Overclaim

28%

True ROAS

3.0x

Based on cross-brand holdout testing benchmarks. Actual overclaim varies by vertical, audience, and campaign type.

PlatformAvg OverclaimRangePrimary DriverWorst Category
Meta Ads28%15-45%View-through attributionRetargeting campaigns
Google Ads18%8-30%Branded search cannibalizationBrand campaigns
TikTok Ads35%20-55%View-through + broad attributionAwareness campaigns
Pinterest Ads22%12-35%View-through windowsHome & lifestyle
Snap Ads30%18-45%View attribution defaultsYounger demographics
TikTok has the highest average overclaim (35%) because its content format - autoplay video - generates massive view-through attribution even when users aren't paying attention. Meta follows at 28%, primarily from view-through and organic cannibalization. Google is lowest at 18%, but branded search cannibalization means the true incremental value of Google Brand campaigns is often near zero.

Chapter 5: Holdout Testing - The Gold Standard

The most reliable way to measure true attribution is to stop showing ads to a subset of your audience and measure the difference. This is holdout testing - the gold standard of incrementality measurement.

How to Run a Holdout Test

1

Define your holdout

Select 10-20% of your audience (by geo, cohort, or random split) to receive zero ads from the channel you're testing.

2

Run for 2-4 weeks

The test needs enough time to capture full purchase cycles. For higher-AOV products, run longer.

3

Measure the delta

Compare conversion rates between the exposed group and holdout group. The difference is your true incremental lift.

4

Calculate true ROAS

Incremental revenue (exposed - holdout) ÷ ad spend = true incremental ROAS. This is always lower than platform-reported.

5

Apply correction factor

Platform-reported ROAS ÷ true ROAS = your correction factor. Apply this to all future platform data.

Budget Consideration

Holdout testing means deliberately not showing ads to some potential customers. For a brand spending $100K/month on Meta, a 15% holdout means ~$15K of “foregone” impressions for 3 weeks. The short-term cost is real, but the long-term value of accurate attribution data saves multiples of that amount in misallocated spend.

Chapter 6: Building Your Attribution Model

You don't need a data science team to build a reliable attribution model. Here's the practical framework:

Step 1: Baseline

Run holdout tests on your top 2-3 channels to establish correction factors. Start with your biggest spend channels - the overclaim there costs the most money.

Step 2: Correct

Apply correction factors to all platform-reported data. If Meta overclaims by 28%, multiply all Meta ROAS by 0.72 to get corrected numbers.

Step 3: Unify

Create a single source of truth combining corrected platform data with Shopify/revenue data. This is your de-biased view.

Step 4: Iterate

Re-run holdout tests quarterly. Overclaim rates change with audience saturation, creative mix, and platform algorithm updates.

The key insight: you don't need perfect attribution. You need attribution that's directionally correct enough to make better allocation decisions. Even a rough correction factor - knowing Meta overclaims by “roughly 25-30%” - dramatically improves your budget decisions compared to trusting raw platform numbers.

Chapter 7: Platform-Specific Correction Factors

Based on holdout testing across 100+ ecommerce brands, here are the correction factors by platform and campaign type. Apply these to platform-reported ROAS to estimate true incremental ROAS.

PlatformCampaign TypeCorrection FactorExample
MetaProspecting (broad)0.68-0.78Reported 3.5x → True ~2.5x
MetaRetargeting0.45-0.60Reported 8.0x → True ~4.2x
MetaAdvantage+ Shopping0.70-0.80Reported 4.0x → True ~3.0x
GoogleNon-brand Search0.80-0.90Reported 5.0x → True ~4.3x
GoogleBranded Search0.15-0.35Reported 12x → True ~3.0x
GooglePerformance Max0.65-0.75Reported 4.5x → True ~3.2x
TikTokSpark Ads0.60-0.75Reported 3.0x → True ~2.0x
TikTokIn-Feed Video0.55-0.70Reported 2.5x → True ~1.6x
The most shocking number: branded search typically has only 15-35% true incrementality. Those “12x ROAS” branded campaigns? The vast majority of those customers would have found you anyway. This is the single biggest misallocation we see across ecommerce brands - over-investing in branded search because the reported ROAS looks amazing.

Chapter 8: Automating Attribution with Parker

Everything in this guide is what Parker - Cresva's attribution agent - executes automatically. Parker continuously monitors platform-reported data, applies correction factors derived from ongoing holdout testing, and feeds de-biased numbers to Felix (forecasting) and Sam (strategy).

What Parker Does, Continuously

Monitors platform-reported ROAS across Meta, Google, TikTok, and Pinterest

Applies and updates correction factors based on ongoing holdout calibration

Detects when overclaim rates change (seasonal shifts, algorithm updates, audience saturation)

Feeds corrected numbers to Felix for forecasting and Sam for budget allocation

Alerts you when a channel's true incremental ROAS drops below your target

Generates attribution reports comparing platform-reported vs corrected performance

The result: you never make a budget decision based on inflated platform numbers again. Every ROAS figure you see in Cresva has been corrected for platform overclaim, giving you the true incremental picture.

This entire methodology is what Parker runs 24/7 on your data. No manual holdout tests. No spreadsheet correction factors. No hoping the platforms are honest. Just accurate attribution, continuously updated, feeding every other decision in the system.

Written by the Cresva Team

Questions? Email us