Central Limit Theorem and Sampling Distributions
The Central Limit Theorem and sampling distributions form the backbone of inferential statistics, essential for AP Statistics, college stats courses, and the GRE Quantitative section. Understanding how sample means behave — and why they approximate a normal distribution — unlocks confidence intervals, hypothesis tests, and virtually every statistical inference procedure.
Interactive Deck
5 CardsWhen does the CLT apply?
How does sample size affect sampling distribution?
Master this topic effortlessly.
Study G helps you master any topic effortlessly using proven learning algorithms and smart review timing
Download Study GFrequently Asked Questions
What is the difference between a population distribution and a sampling distribution?
Population distribution describes all individual values in a population, while a sampling distribution describes the distribution of a statistic computed from repeated samples.
- Population distribution: values of individuals
- Sampling distribution: values of sample statistics
Why is the Central Limit Theorem important in statistics?
The CLT is the foundation of inferential statistics. It guarantees that, for large enough samples, the distribution of sample means is approximately normal — enabling the use of z-tests, t-tests, and confidence intervals even when the population is not normal.
How large must a sample be for the CLT to apply?
A sample size of n ≥ 30 is the standard rule of thumb for the CLT to produce an approximately normal sampling distribution. However, larger samples are needed for heavily skewed or multi-modal populations.
