Skip to main content
Market Intelligence Wiki

Sample Data Verification

Last updated May 2026

Definition

Defines sample data verification as the procurement-stage discipline of evaluating market intelligence datasets before purchase — what to ask for, what to test, and what disqualifies a vendor.

Sample data verification is the discipline of evaluating a market intelligence dataset against procurement-stage tests before purchase, rather than discovering its limitations after deployment. It is the single most leveraged decision-quality activity in market intelligence procurement — twenty minutes of stress-testing on a sample regularly avoids six-figure bad bets on unsuitable datasets.

It is also, in practice, the activity most often skipped. Procurement teams default to evaluating vendors on demo polish, sales-process responsiveness, and feature-list parity. None of those correlate well with dataset quality.

Why demos are not enough#

A vendor demo shows the dashboard. The dashboard is the rendering layer, not the data layer. The two are designed to look good together — demos use curated, complete categories where the underlying coverage is densest. The sparse-coverage categories that will frustrate the buyer post-purchase rarely make it into the demo deck.

A sample dataset, by contrast, surfaces:

  • Coverage gaps — which platforms have thin data, which date ranges have collection failures, which categories have under twenty SKUs of depth
  • Classification errors — sentiment misreads, attribute extraction misses, theme misgrouping
  • Methodology shortcuts — bot inclusion in social counts, repost double-counting, language-detection failures
  • Normalisation issues — currency conversion approaches, platform-specific volume reconciliation, cross-platform de-duplication

The dashboard hides all of these. The sample dataset reveals all of them. Buyers who purchase on demo polish without a sample data check are essentially purchasing the dashboard, not the data.

What a defensible sample looks like#

Three properties separate a useful sample from a marketing artefact:

Representative scope. Covers the actual platforms, categories, and time window the buyer plans to use. A skincare-buyer is not served by a sample that demonstrates only consumer electronics. A China-market buyer is not served by a sample that covers only the United States. Buyers should specify the scope; vendors that cannot match the scope at sample stage usually cannot match it at production stage either.

Audit-ready depth. Lets the buyer drill from any aggregate number back to the underlying source records. A sample showing "23% negative sentiment on packaging across 8,400 reviews" should let an analyst pull the actual reviews making that classification. Without drill-down, the aggregate is unverifiable, and unverifiable aggregates have no place in million-dollar decisions.

Methodology-paired. Comes with a written description of collection, exclusions, and normalisation. A sample without methodology documentation invites the buyer to assume best-case behaviour where the vendor may be doing worst-case. The documentation should describe how data is sourced, what is excluded and why, and how cross-platform figures are reconciled. Pairs with the broader How to Evaluate Market Intelligence Providers criteria.

Tests an analyst can run in an afternoon#

Four tests filter most unsuitable datasets without requiring statistical sophistication:

1. SKU-completeness check. Pick five known top-selling SKUs in the buyer's category. Search the sample for each. All five should appear with reasonable depth (review counts, social mentions, sales-derived metrics). If two or more are missing or under-covered, the dataset has coverage gaps in the category the buyer cares most about.

2. Date-range consistency. Plot record volume by week or month across the claimed window. Coverage should be approximately uniform, with seasonal variation matching expected category patterns. Gaps that look like collection failures (sudden drop to near-zero followed by recovery) indicate operational fragility — the same gaps will appear in production.

3. Sentiment sanity check. Pull fifty randomly-sampled negative-classified reviews. Read each. The buyer's read should agree with the classification on at least eighty percent of cases — if agreement is below seventy percent, the sentiment model is not category-tuned. Repeat with positive-classified reviews. The neutral fraction should be under twenty percent (see Sentiment Analysis for why).

4. Methodology-concordance check. Read the methodology document. Verify three claims in the document against the sample data. If the methodology says "duplicate reviews are removed", spot-check for duplicates. If it says "bot accounts are filtered", spot-check for accounts with obvious bot signatures. Concordance failures are usually the most diagnostic — they reveal where the vendor's marketing language diverges from operational practice.

When a vendor refuses to provide sample data#

Common reasons, in order of severity:

  1. Licensing constraints from upstream data sources. Legitimate but should be disclosed and worked around (curated samples, on-site analyst review, NDA-protected access).
  2. Proprietary methodology concerns. Legitimate but solvable — methodology can be described at a process level without exposing trade-secret implementation.
  3. Thin coverage in the requested category. The vendor knows the sample would expose the gap. This is the most common reason and the most damaging signal.
  4. No process for providing samples. Indicates the vendor has not been asked before — which means most existing customers bought without verification, which means existing customers are a worse reference than they appear to be.

A vendor refusal without a clear reason is a strong negative signal. Verifiability is a baseline requirement for intelligence-grade datasets; a vendor uncomfortable with it is selling a different category of product.

How sample data verification fits the broader procurement process#

Sample data verification is one criterion in the broader How to Evaluate Market Intelligence Providers framework. It interacts with methodology transparency (a methodology document without a sample is unverifiable), historical depth (claimed coverage windows can be tested at the date-consistency level), and use-case fit (sample-tested categories are the only ones with confirmed fit).

Run it before signing. The cost of running it is a half-day of analyst time. The cost of skipping it is the difference between the dataset the buyer thought they were buying and the dataset they actually got.

Where to look next#

For the full procurement framework, see How to Evaluate Market Intelligence Providers. For the methodological substrate that makes verification possible, see Sentiment Analysis. For the discipline-level overview, see Market Intelligence Overview.

Common questions#

Should you trust a vendor demo or insist on a sample dataset?#

Demos show the dashboard. Sample datasets show the data. The two reveal different things: a polished demo can hide thin underlying coverage, while a sample dataset surfaces gaps, classification errors, and methodology shortcuts that a demo papers over. Procurement teams that purchase on demo alone consistently report dataset disappointment within ninety days. Insisting on a sample dataset before purchase — ideally for the specific category, platforms, and time window the buyer cares about — is the cheapest insurance available.

What does a defensible sample dataset look like?#

Three properties: representative scope (covers the actual platforms, categories, and time window the buyer will use), audit-ready depth (lets the buyer drill from any aggregate number back to the underlying source records), and methodology-paired (comes with a written description of collection, exclusions, and normalisation). A sample that is none of these is a marketing artefact. A sample that is all three lets a procurement analyst stress-test the dataset against questions the vendor was not given in advance — which is the only meaningful test.

What are the cheapest tests to run on a sample dataset?#

Four tests catch most quality issues in a single afternoon: SKU-completeness check (pick five known top-sellers in the category, verify all five appear in the data), date-range consistency (confirm coverage is uniform across the claimed window with no large gaps), sentiment sanity check (read fifty randomly-sampled classifications and compare to your own read), and methodology-concordance check (read the methodology document and verify the sample matches what was described). None require statistical sophistication; all four together filter ninety percent of unsuitable datasets.

When does a vendor refuse to provide sample data?#

Common reasons: thin coverage in the category the buyer is asking about (which the sample would surface), proprietary methodology that the vendor does not want competitors to see (legitimate, but should still be solvable with NDAs and curated samples), or licensing constraints from the vendor's own data sources (also legitimate, but should be disclosed). A vendor that refuses without a clear reason is a strong negative signal — verifiability is a baseline requirement for intelligence-grade datasets, and a vendor uncomfortable with it is selling a different category of product.

Talk to a Market Intelligence Specialist

Get a sample of our market intelligence data — covering your category, your platforms, your markets.

By submitting, you agree to our Privacy Policy .