Sampling Distributions
A sampling distribution describes how a sample statistic (such as the sample mean) varies from sample to sample. The Central Limit Theorem is the cornerstone result, stating that the distribution of sample means approaches a normal distribution as sample size increases, regardless of the population's shape. Understanding sampling distributions bridges descriptive statistics and inferential statistics.
Solve Sampling Distributions Problems with AI
Snap a photo of any sampling distributions problem and get instant step-by-step solutions.
Download StatsIQKey Concepts
Study Tips
- โDistinguish clearly between the population distribution, the sample distribution, and the sampling distribution of a statistic. These are three different things and confusing them leads to errors.
- โRemember the CLT conditions: the sample size should generally be at least 30, or the population should be approximately normal. For proportions, check that np >= 10 and n(1-p) >= 10.
- โThe standard error shrinks by a factor of the square root of n. To cut the standard error in half, you need to quadruple the sample size, not double it.
- โSimulate sampling distributions using software. Drawing thousands of samples and plotting their means makes the CLT tangible and helps you see the normal shape emerge.
Common Mistakes to Avoid
Students frequently confuse standard deviation with standard error. Standard deviation measures spread in the data, while standard error measures spread of a sample statistic across repeated samples. Another common error is applying the CLT when the sample size is too small and the population is heavily skewed. Students also sometimes believe the CLT says the data become normally distributed with large n, when in fact it is the distribution of the sample mean (or sum) that becomes normal. The underlying data distribution does not change.
Sampling Distributions FAQs
Common questions about sampling distributions
The Central Limit Theorem states that the sampling distribution of the sample mean is approximately normal when the sample size is sufficiently large, regardless of the shape of the population distribution. This matters because it justifies using normal-based inference procedures (z-tests, confidence intervals) even when the population is not normal. It is the reason so many statistical methods work in practice. The approximation improves as n increases, and n >= 30 is a common rule of thumb.
Standard deviation (SD) measures the spread of individual data points around the mean within a single sample or population. Standard error (SE) measures the spread of a sample statistic (like the sample mean) across many hypothetical repeated samples. SE = SD / sqrt(n). As sample size increases, individual data points do not become less variable (SD stays the same), but the sample mean becomes a more precise estimate of the population mean (SE decreases).