Skip to content

Central Limit Theorem

In probability theory, the central limit theorem (CLT) states that, under appropriate conditions, the distribution of a normalized version of the sample mean converges to a standard normal distribution. This holds even if the original variables themselves are not normally distributed. There are several versions of the CLT, each applying in the context of different conditions.

Some of the important versions of this are:

VariantAssumptionsFormula TypeUse Case
Classical / Lindeberg–Lévyi.i.d., finite varianceZnN(0,1)General inference, sample mean
LyapunovIndependent, not identically dist.Normalized sum N(0,1)Heterogeneous data
Lindeberg–FellerIndependent, with Lindeberg conditionGeneral normalized sumMost general CLT
Multivariate CLTi.i.d. vectorsn(X¯μ)Nd(0,Σ)Multivariate stats, finance
Functional CLTi.i.d., viewed as stochastic processSn(t)W(t)Time series, Brownian motion modeling
Dependent Variable CLTWeakly dependent varsSum N(0,1)Time series, Markov chains

Classical Central Limit Theorem

If X1,X2,...,Xn are independent and identically distributed (i.i.d.) random variables with:

  • Mean μ,
  • Variance σ2<,

Then the standardized sum:

Zn=i=1nXinμσndN(0,1)

as n.

Application 1 - Usage of Standard Deviation(σ)

CLT says:

The sampling distribution of the sample mean ( x¯ ) will be approximately normal with:

  • Mean: μ (same as the population mean)

  • Standard Deviation (of the sample mean):

    σx¯=σn

Where:

  • σx¯ = standard deviation of the sampling distribution (also called standard error)
  • σ = population standard deviation
  • n = sample size

How σ Works with CLT

1. Reduces Variability with Bigger Samples

As sample size n increases:

σx¯=σngets smaller

This means: Your sample mean becomes more stable and reliable as you take larger samples.

2. Allows Confidence Interval Estimation

In business, we often construct a confidence interval for the mean:

x¯±zσn

Where:

  • x¯ = sample mean
  • z = z-score (e.g., 1.96 for 95% confidence)
  • σ/n = standard error

You need σ to determine how “wide” or “narrow” your estimate is.

3. Enables Hypothesis Testing

When comparing two groups (A/B testing, quality control, etc.), you calculate:

Z=x¯μσ/n
  • A smaller σ → more precise Z-scores → clearer decisions.

Example

Let’s say a factory produces nails with a known population standard deviation σ=2 mm, and you take a sample of 100 nails.

  • Standard error:

    σx¯=2100=0.2

So your sampling distribution of the mean length will have:

  • Mean = population mean (say, 50 mm)
  • Std dev = 0.2 mm, not 2 mm!

Your estimate of the average length from the sample is much more precise than any one nail.

Summary

ConceptRole of σ (sigma)
Sampling distributionDetermines spread of sample means
Standard errorσ/n: lower with bigger n
Confidence intervalsUsed to calculate margin of error
Hypothesis testing (Z-tests)Critical in test statistic formula

If the population σ is unknown, we use the sample standard deviation s and apply the t-distribution — another extension of CLT.

Powered by VitePress