| Study Guides
AP Statistics · Sampling Distributions · 14 min read · Updated 2026-05-09

Sampling Distributions — AP Statistics Study Guide

For: AP Statistics candidates sitting AP Statistics.

Covers: Sampling distribution of a sample mean and a sample proportion, Central Limit Theorem, conditions (Random / 10% / Normal or Large Counts) for using normal approximations — AP Statistics Unit 5.

You should already know: Probability and random variables (Unit 4), normal distribution (Unit 1).

A note on the practice questions: All worked questions in the "Practice Questions" section below are original problems written by us in the AP Statistics style for educational use. They are not reproductions of past College Board papers. Use them to practise the technique; cross-check with official College Board mark schemes for grading conventions.


1. Why Sampling Distributions Matter

Unit 5 is the conceptual bridge between probability (Units 3-4) and inference (Units 6-9). Every confidence interval and hypothesis test you'll learn from Unit 6 onwards rests on knowing the sampling distribution of the relevant statistic. About 7-12% of the AP Stats score comes from Unit 5 directly, but its concepts power the remaining 30+%.

The single big idea: a statistic computed from a random sample is itself a random variable, with its own probability distribution called the sampling distribution. Knowing the shape, centre, and spread of this distribution lets you say "the sample mean is unlikely to be more than X away from the population mean".

2. Sampling distribution of a sample proportion

If we take samples of size from a population with true proportion , the sample proportion has:

  • Mean: (the sampling distribution is unbiased).
  • Standard deviation (= standard error): .
  • Shape: approximately Normal when Large Counts condition holds: AND .

Conditions for using these formulas:

  1. Random: data come from a random sample.
  2. 10% condition: (sample is no more than 10% of population, so independence of draws is approximately valid).
  3. Large Counts: as above.

3. Sampling distribution of a sample mean

For samples of size from a population with mean and standard deviation :

  • Mean: .
  • Standard deviation: (decreases as ).
  • Shape: approximately Normal when either the population is Normal or (Central Limit Theorem).

Conditions:

  1. Random: random sample.
  2. 10% condition: .
  3. Normal or large : as above (CLT).

4. Central Limit Theorem (CLT)

For a sample of size drawn from any population (Normal, skewed, bimodal, etc.) with finite mean and standard deviation :

The sampling distribution of approaches Normal(, ) as .

Practical rule of thumb: is "large enough" for most populations. Exceptions: extremely skewed populations may need larger for the approximation to be good.

CLT does not say:

  • The original population becomes Normal (it doesn't).
  • Individual data points are Normally distributed (they're not).
  • The shape of for small is Normal (it isn't, unless population is).

5. Bias and variability

A statistic is unbiased if its mean equals the parameter it estimates (e.g. ). and are both unbiased.

Variability is the spread of the sampling distribution — controlled by (larger → less variable). To halve the standard error, multiply by 4.

A good estimator is both unbiased and low-variability. A precise but biased estimator (always 5% off) and a high-variability unbiased estimator (averaging right but each estimate wildly different) are both bad.

6. Differences and combinations

Sampling distribution of (independent samples):

Sampling distribution of (independent samples):

Variances add (under independence); standard deviations don't. This is the foundation of two-sample inference.

7. Worked Example

A factory claims 92% of its widgets pass inspection. A consumer group samples widgets randomly and finds 175 pass (so ). What is the probability of getting a sample proportion of 0.875 or lower if the true rate is really 92%?

Solution.

  1. Conditions check:

    • Random ✓ (stated).
    • 10% condition: assume , i.e. factory makes ≥ 2000 widgets — reasonable.
    • Large Counts: ✓; ✓.
  2. Sampling distribution of : approximately Normal with , .

  3. Z-score: .

  4. Probability: .

Interpretation: only ~1% chance of seeing a sample proportion this low if the company's claim is true. That's strong evidence against the 92% claim — suggests the true rate is below 92%.

8. Common Pitfalls

  • Confusing vs : the population SD is one thing; the SD of the sampling distribution of the mean is , which is smaller. AP graders catch this often.
  • Ignoring conditions: state and check Random, 10%, and Normal/Large Counts conditions every time on the FRQ. Missing one loses points.
  • CLT confusion: CLT applies to the sampling distribution, not the population. The population's shape stays whatever it is.
  • Sample size confusion: bigger → smaller standard error, not smaller population SD.

9. Practice Questions (CED Style)

  1. A population has mean and SD . For samples of size 36, calculate and find . Justify the use of Normal approximation.
  2. A poll surveys 1000 voters. The true population proportion supporting candidate A is 0.55. Find the probability the sample proportion will be between 0.52 and 0.58.
  3. Two independent factories produce circuit boards. Factory A averages 5 defects per 100 boards (n=400), Factory B averages 3 defects per 100 boards (n=300). Find the standard error of , and the probability the observed difference is greater than 4 percentage points.

10. Quick Reference Cheatsheet

  • : , , Normal if AND .
  • : , , Normal if pop. Normal OR .
  • Conditions: Random, 10%, Normal/Large Counts. State all 3 every time.
  • CLT: approaches Normal as for any population shape.
  • Difference of two: variances add, .
  • Halve SE: quadruple .

11. What's Next

Sampling distributions are the foundation for Unit 6 (Inference: Proportions) and Unit 7 (Inference: Means) — confidence intervals and hypothesis tests are direct applications. Drill the conditions check until it's automatic. Use Ollie to step through specific FRQ scenarios: "Why is my sample size too small to use the Normal approximation here?" or "Walk me through the Z-score logic for a difference of two means".

← Back to topic

Stuck on a specific question?
Snap a photo or paste your problem — Ollie (our AI tutor) walks through it step-by-step with diagrams.
Try Ollie free →