How-to

How to Determine Survey Sample Size: Statistical Basis and Practical Guidelines

How to determine the right survey sample size — from Cochran's formula and confidence intervals to industry rules of thumb, plus what to know about running on small samples.

"How many responses do I actually need for this survey?" If you've ever run research, you've either asked this question or been asked it. The honest answer is "there's a well-defined statistical formula, but real-world sample size decisions involve more than just the math." Plenty of projects get launched on "let's go with 100" or "whatever the budget allows," then hit the review stage and get pushed back for inadequate sample size. A common, quiet failure mode.

This piece covers the statistical basis for sizing a survey (Cochran's formula, confidence level, margin of error), the industry rules of thumb that vendors use in practice, and the risks of running on small samples. The goal: you walk into your next sizing conversation with a defensible answer, not a guess.

1. What Sample Size Actually Controls

Before the math, a clear picture of what sample size is trading off:

  • Confidence level: if you ran the same survey 100 times, how often would the result land inside the target range. 95% is the industry default
  • Margin of error: the acceptable gap between your sample result and the "true" population value. ±5% is standard
  • Representativeness: how well your sample reflects the underlying population's distribution

More sample → higher confidence, narrower margin, better representativeness. It also costs proportionally more, so the decision is where additional sample stops being worth it. That's where the statistics come in.

2. The Statistical Basis — Cochran's Formula

The classical formula for sample size determination is Cochran's Formula, introduced by William G. Cochran in 1977. It remains the reference in textbooks and practitioner guides as the standard methodology for this decision.

The formula

For an infinite (or unknown) population:

n0=Z2p(1p)e2n_0 = \frac{Z^2 \cdot p(1 - p)}{e^2}

Where:

  • n0n_0: required sample size
  • ZZ: Z-score for the chosen confidence level (Z=1.96Z = 1.96 for 95%, Z=2.58Z = 2.58 for 99%)
  • pp: assumed population proportion. Use p=0.5p = 0.5 when unknown — it's the most conservative choice (produces the largest sample estimate)
  • ee: margin of error (typically e=0.05e = 0.05 = ±5%)

Plugging in 95% confidence (Z=1.96Z = 1.96), p=0.5p = 0.5, e=0.05e = 0.05:

n0=1.962×0.5×0.50.052=384.16n_0 = \frac{1.96^2 \times 0.5 \times 0.5}{0.05^2} = 384.16

About 384 responses suffice for 95% confidence at ±5% error, for an infinite population.

Finite population correction

If your population is finite — say, your 1,000 customers — you apply the Finite Population Correction (FPC):

n=n01+n01Nn = \frac{n_0}{1 + \dfrac{n_0 - 1}{N}}

Where NN is the population size. Smaller populations need proportionally fewer responses.

Canonical sample size table

With 95% confidence, ±5% error, p = 0.5:

Population (N)Required sample (n)
10080
500217
1,000278
10,000370
100,000+ / infinite384

For the full derivation, see Cochran's formula explainer at Statistics How To and sample size determination in applied research on ResearchGate.

3. Industry Rules of Thumb

Cochran's formula gives a statistical minimum. In practice, other factors also drive sample size decisions, so vendor guides publish rules of thumb. These should be treated as widely shared industry reference values rather than rigorously validated benchmarks.

Error ranges by sample size

Across vendor commentary, the following ranges are commonly cited:

SampleApproximate margin of errorPractical interpretation
50~±15%Rough directional read only
100~±10%Reasonable reliability
400~±5%Standard research-grade precision
1,000~±3%High precision, detects small differences

The industry convention of "N=400 is the baseline" traces to the fact that ~384 is the statistical minimum for ±5% error at 95% confidence.

Cost-precision balance — why "around 400 is the sweet spot"

When you plot this relationship, the curve has a clear steep early section and a flat tail. One picture makes the whole tradeoff obvious.

N=400 · ±4.9%Best cost-precision balanceN=100 · ±9.8%N=1000 · ±3.1%0400800120016002000Sample size0%4%8%12%16%Margin of error (±%)
Calculated at 95% confidence (Z=1.96), p=0.5

Going from N=100 to N=400, margin of error drops from about 9.8% to 4.9% — a halving. Going from N=400 to N=1,000, it only drops from 4.9% to 3.1% — just 1.8 points for 2.5× the sample. This comes from n\sqrt{n} in the formula e=1.96p(1p)/ne = 1.96 \sqrt{p(1-p)/n}: to halve the margin of error, you need to quadruple the sample.

The industry convention of N=400 isn't arbitrary — it sits right at the point where precision gains start flattening out. Past that point, each additional respondent buys much less precision per dollar. Around 400 is where the price-precision trade-off genuinely peaks.

Vendor sample size calculators

Multiple vendors host free online calculators with the same input fields (population, confidence, margin):

All of them implement Cochran's formula under the hood, so results are effectively identical. The math is solved; the human judgment is in the inputs — what confidence level and what margin of error do you actually need?

4. Sample Size by Use Case

Different study types need different precision. Common industry patterns:

Exploratory / screening studies

  • Goal: Pre-hypothesis directional read
  • Range: 30–100 responses
  • When it's OK: When ±10–15% error won't change the decision

Decision-grade studies

  • Goal: Product, service, or executive-level decisions
  • Range: 300–500 responses total
  • When it applies: Standard research-grade at 95% confidence, ±5%

Segment comparison studies

  • Goal: Compare across demographics, regions, job roles, etc.
  • Range: At least 100 per segment, with totals often 800–2,000+
  • Why: Margin of error within each segment also needs to be ±10% or tighter

Longitudinal / tracking studies

  • Goal: Repeat the same survey over time and track change
  • Range: 400–1,000 per wave
  • Why: Need statistical power to detect wave-over-wave differences

The worst antipattern is deciding the sample number before the purpose. Purpose first, number second.

5. What Happens with Small Samples

Sometimes the budget only allows N=50. Understanding exactly what that costs you lets you make an informed decision.

Risk 1: Wide confidence intervals

At N=50, margin of error is around ±15%. A result of "40% purchase intent" actually means "somewhere in the 25–55% range." That's often too wide to support a firm claim one way or the other.

Risk 2: Segmentation becomes meaningless

Split N=50 into male/female (25/25), then by age bracket — each cell ends up at 5–10 responses. Statistical comparison between segments of that size is essentially noise.

Risk 3: Outliers dominate the signal

One extreme respondent moves the whole score materially. In open-text coding, minority views can appear as "the majority opinion" simply because the majority is small.

When small samples are legitimately fine

  • Qualitative insight generation: "What kinds of complaints exist?" works fine at N=30
  • Hypothesis validation before a bigger study: pilot studies
  • Targeted supplementation: you already have large-scale data and want a small deep-dive

The rule at small sample sizes: don't let the numbers walk on their own. Always state the margin.

6. Editorial Take — Four Sample Size Mistakes We See Repeatedly

After tracking published case studies and industry commentary over time, four sample size failures come up again and again. We'll call them out directly.

1. "Let's just do 100" is riskier than people realize. At N=100, margin of error is ±10%. A "40% purchase intent" result could easily be anywhere from 30% to 50% — a range wide enough to completely change a decision. If you ship at 100, spell out the margin in the report. Reports that quietly present N=100 results without uncertainty bands are the ones that later get challenged as "not rigorous enough."

2. Setting total sample size while ignoring segment cells. "We'll collect 500 total" is fine, right up until you open the data and find your critical segments have 30 responses each. If segmentation is part of the analysis, size up from the smallest cell. Want to compare 40–59-year-old women's Top-2 Box scores? Set the minimum segment at 100 first; total size follows from that.

3. Confusing "sample needed" with "sends needed." We still see people target 384 sends when they need 384 responses. This is always wrong. If your email response rate is 10%, you need to send 3,840 — not 384. Factor in response rate up front. See the response-rate fundamentals for more.

4. "Bigger sample = more accurate" is a common but wrong intuition. Sample size controls precision, but a biased sample is still biased at any size. 10,000 responses from your heaviest users is not a representative study. Representativeness matters more than raw volume. A well-targeted 300-response study usually beats a 10,000-response convenience sample for actual decision-making.

7. Designing Surveys in the Survey Tool Kicue

Kicue includes the features you'll need to execute on sample size decisions:

  • Quota management — set target samples per segment and auto-close when targets are hit (details)
  • Real-time collection monitoring — track segment-level progress in a live dashboard (details)
  • URL parameter routing — receive attributes from external panels to target distribution precisely
  • Response-rate features — mobile optimization, skip logic, progress bar to maximize completion against sends

Upload a questionnaire file and Kicue handles the distribution, quota, and analysis plumbing you'd otherwise be stitching together.

Choosing the right tool — Free plan limits, branching support, AI capabilities, and CSV export vary widely across tools. See our free survey tool comparison to find the right fit for this approach.

Recap

A checklist for sample size decisions:

  1. Cochran's formula is the statistical baseline — 95% confidence at ±5% error ≈ 384 for infinite populations
  2. Apply finite population correction — N=100 population needs 80, N=1,000 needs 278
  3. Size from the use case — screening 30–100, decision-grade 300–500, segment comparison 100+ per cell
  4. Know what small samples cost you — ±15% margins, no segment analysis, outlier sensitivity
  5. Representativeness beats volume — a biased sample at 10,000 doesn't beat a representative sample at 300

The question "how many do we need?" is answered by working backward from purpose and required precision. Once that becomes a habit, you can defend your sample size in any meeting — including the one where the budget is being negotiated.


References (8)

Handle sample size, quota management, and distribution in one place with Kicue — a free survey tool that ships with quota targeting and live monitoring.

Related articles

Ready to create your own survey?

Upload your survey file and AI generates a web survey form in 30 seconds.

Get started for free