Why This Concept Exists
Sampling distributions answer a fundamental question: what is the probability distribution of a statistic? Every confidence interval, every hypothesis test, every p-value in PTS2 depends on knowing the sampling distribution of a relevant statistic. Without this machinery, inference is impossible.
The three cornerstone distributions — chi-squared, t, and F — are all constructed from standard normal random variables and have deep interconnections. Every formula you need in N8 through N12 traces back to how these distributions are built. If you don't internalise how \(\chi^2\) arises from squaring normals, how t arises from a normal divided by a chi, and how F arises from a ratio of chi-squares, then every confidence interval formula in later nodes will feel like unmotivated memorisation.
This node also introduces two powerful techniques that appear in exam questions: constructing statistics from heterogeneous normal samples (different means and variances, standardising them, and building chi-squared/t/F statistics from the result), and understanding additive properties of chi-squared and F distributions.
Prerequisites
Before engaging with this node, you must be comfortable with:
- Standard Normal distribution: \(Z \sim N(0,1)\) with PDF \(\phi(z) = \frac{1}{\sqrt{2\pi}}e^{-z^2/2}\). Know that if \(X \sim N(\mu, \sigma^2)\) then \(Z = \frac{X-\mu}{\sigma} \sim N(0,1)\).
- Linear combinations of normals: If \(X_i \sim N(\mu_i, \sigma_i^2)\) are independent, then \(\sum a_i X_i \sim N(\sum a_i\mu_i, \sum a_i^2\sigma_i^2)\).
- Moment generating functions (MGFs): The technique of using \(M_Z(t) = E[e^{tZ}]\) to identify distributions. If two variables have the same MGF, they have the same distribution.
- Independence: The sample mean \(\bar{X}\) and sample variance \(S^2\) are independent for normal samples. This is a deep result (Basu's theorem) that must be treated as a black box at the PTS2 level.
- Gamma function: \(\Gamma(\alpha) = \int_0^{\infty} x^{\alpha-1}e^{-x}\,dx\), and \(\Gamma(n) = (n-1)!\) for positive integers.
- Sample variance formula: \(S^2 = \frac{1}{n-1}\sum_{i=1}^{n}(X_i - \bar{X})^2\), with Bessel's correction (\(n-1\) denominator).
Core Exposition
3.1 The Chi-Squared Distribution — Construction
The chi-squared distribution is the sum of squares of independent standard normals:
\(Y = \displaystyle\sum_{i=1}^{\nu} Z_i^2 \sim \chi^2(\nu)\)
\(\nu\) is called the degrees of freedom.
The PDF of \(Y \sim \chi^2(\nu)\) is:
Key properties:
MGF: \(M(t) = (1 - 2t)^{-\nu/2}\) for \(t < 1/2\)
This is the bridge between algebraic computation (sample variance) and probability theory (chi-squared distribution). The proof uses Cochran's theorem and is far beyond PTS2 scope, but the result is fundamental.
3.2 Additivity of Chi-Squared
\(X_1 + X_2 \sim \chi^2(\nu_1 + \nu_2)\).
Proof via MGFs: \(M_{X_1+X_2}(t) = (1-2t)^{-\nu_1/2} \cdot (1-2t)^{-\nu_2/2} = (1-2t)^{-(\nu_1+\nu_2)/2}\). \(\checkmark\)
3.3 The t-Distribution — Construction
The t-distribution arises when dividing a standard normal by the scaled square root of an independent chi-squared:
\(T = \dfrac{Z}{\sqrt{V/\nu}} \sim t(\nu)\)
The PDF of \(T \sim t(\nu)\) is:
The t-distribution is symmetric about zero, bell-shaped like the normal, but with heavier tails. As \(\nu \to \infty\), \(t(\nu) \to N(0,1)\).
3.4 The F-Distribution — Construction
The F-distribution arises as a ratio of scaled independent chi-squared variables:
\(F = \dfrac{U/\nu_1}{V/\nu_2} \sim F(\nu_1, \nu_2)\)
Note that each chi-squared is divided by its degrees of freedom before forming the ratio. The PDF is complex, but the structure is what matters for exams.
Key relationships:
\(F_{\alpha;\,\nu_1,\nu_2} = \dfrac{1}{F_{1-\alpha;\,\nu_2,\nu_1}}\) (reciprocal property for critical values).
3.5 Constructing Statistics from Heterogeneous Normals
A common exam pattern: you are given \(X_1, \ldots, X_n\) where each \(X_i \sim N(\mu_i, \sigma_i^2)\) with potentially different means and variances. You must standardise each individually and then combine them to produce a chi-squared, t, or F statistic.
\(\displaystyle \sum_{i=1}^{n} Z_i^2 = \sum_{i=1}^{n}\left(\dfrac{X_i - \mu_i}{\sigma_i}\right)^2 \sim \chi^2(n)\).
If \(\mu_i\) are unknown and replaced by the sample mean \(\bar{X}\), one degree of freedom is lost:
3.6 Key Sampling Distributions for Normal Samples
\(\dfrac{\bar{X} - \mu}{\sigma/\sqrt{n}} \sim N(0,1)\)
\(\dfrac{\bar{X} - \mu}{S/\sqrt{n}} \sim t(n-1)\)
\(\dfrac{(n-1)S^2}{\sigma^2} \sim \chi^2(n-1)\)
Worked Examples
Example 1: Building a Chi-Squared Statistic from Different Normals
Let \(X_1, \ldots, X_5\) be independent random variables where \(X_i \sim N(i, i)\) for \(i = 1, 2, \ldots, 5\). Let \(Q = \sum_{i=1}^{5}\left(\dfrac{X_i - i}{\sqrt{i}}\right)^2\).
Find the distribution of \(Q\) and compute \(P(Q > 11.07)\).
The \(Z_i\) are independent because the \(X_i\) are independent.
We have 5 independent \(N(0,1)\) variables squared and summed, so 5 degrees of freedom.
\(P(\chi^2(5) > 11.07) = 0.05\).
Example 2: Constructing a t-Statistic
Let \(X \sim N(3, 4)\) and let \(S^2\) be the sample variance of an independent sample of size 11 from any normal population. Find the distribution of \(T = \dfrac{X - 3}{2 S}\).
But the problem states \(S\) without specifying \(\sigma\). If the underlying population has \(\sigma = 1\) (standardised), then:\(10S^2 \sim \chi^2(10)\).
If \(\sigma = 1\), then \(10S^2 \sim \chi^2(10)\), so \(S = \sqrt{\chi^2(10)/10}\), and:
\(T = \dfrac{Z}{\sqrt{\chi^2(10)/10}} \sim t(10)\).
If the underlying variance is not 1, then we need to rescale accordingly. In practice, exam questions of this type always specify or imply \(\sigma = 1\) for the sample variance population.
Example 3: F-Distribution from Two Independent Sample Variances
Two independent samples are drawn from normal populations with the same variance \(\sigma^2\):
Sample 1: \(n_1 = 5\) observations, sample variance \(S_1^2\).
Sample 2: \(n_2 = 7\) observations, sample variance \(S_2^2\).
Find the distribution of \(F = \dfrac{S_1^2}{S_2^2}\).
Pattern Recognition & Examiner Traps
- "Find the distribution of \(Y = \sum a_i X_i^2\)" — standardise each \(X_i\) individually to \(Z_i\), then the sum is \(\chi^2(\# \text{terms})\). Adjust for constraints (estimated parameters).
- "Show that \(T \sim t(\nu)\)" — explicitly identify the numerator as \(N(0,1)\), the denominator as \(\sqrt{\chi^2(\nu)/\nu}\), and verify independence.
- "Find the probability that \(F > c\)" — recognise as F-distribution, count d.f. in numerator and denominator, look up tables.
- "Find E[T²]" or "Var(F)" — use known formulas for the moments: \(E[t(\nu)^2] = \nu/(\nu-2)\) for \(\nu > 2\), \(E[F(\nu_1,\nu_2)] = \nu_2/(\nu_2 - 2)\) for \(\nu_2 > 2\).
Connections
- ← N4 (Transformations): The derivations of chi-squared, t, and F PDFs from first principles use transformation methods and Jacobian techniques.
- → N7 (Point Estimation): Understanding sampling distributions is essential for evaluating estimators. The MSE of an estimator is computed with respect to its sampling distribution.
- → N8 (Confidence Intervals, One-Sample): The \(t\)-interval uses the t-distribution result. The variance interval uses the chi-squared result. The foundational sampling distributions are used directly in CI construction.
- → N9-N12 (Two-Sample Inference): The F-distribution is used to test equality of variances. Two-sample t-intervals depend on pooled variance having a chi-squared distribution.
In summary: N6 provides the probability distributions, and N8-N12 apply them to inference problems. The link is direct and non-negotiable.
Summary Table
| Distribution | Construction | Key Stats | Support | Key Use |
|---|---|---|---|---|
| \(\chi^2(\nu)\) | \(\sum_{i=1}^\nu Z_i^2, \; Z_i \sim N(0,1)\) | \(E = \nu, \; \text{Var} = 2\nu\) | \((0, \infty)\) | Variance inference |
| \(t(\nu)\) | \(\dfrac{Z}{\sqrt{V/\nu}}\) | \(E = 0 \;(\nu > 1)\) | \((-\infty, \infty)\) | Unknown \(\sigma\) CI |
| \(F(\nu_1,\nu_2)\) | \(\dfrac{U/\nu_1}{V/\nu_2}\) | \(E = \dfrac{\nu_2}{\nu_2-2} \;(\nu_2 > 2)\) | \((0, \infty)\) | Variance ratio test |
Self-Assessment
Test your understanding before moving to N7:
- Construct a \(\chi^2\), \(t\), and \(F\) statistic from given independent normal variables.
- State the PDF, expectation, and variance of each of the three distributions.
- Explain why \(\frac{(n-1)S^2}{\sigma^2} \sim \chi^2(n-1)\) and what each component means.
- Given heterogeneous normals, compute \(\sum\frac{(X_i - \mu_i)^2}{\sigma_i^2}\) and identify its distribution.
- Use the reciprocal property of the F-distribution: \(F_{\alpha;\nu_1,\nu_2} = 1/F_{1-\alpha;\nu_2,\nu_1}\).
- Compute \(E[T^2]\) for \(T \sim t(\nu)\). [Answer: \(\nu/(\nu-2)\) if \(\nu > 2\).]
- If \(X \sim N(0,4)\), what is the distribution of \(X^2/4\)? [Answer: \(\chi^2(1)\).]
- Prove that if \(Y \sim \chi^2(1)\), then \(E[Y] = 1\) and \(\text{Var}(Y) = 2\).
- If \(X_1, X_2, X_3 \sim N(2,1)\), find \(P(\sum (X_i - 2)^2 < 7.815)\). [Answer: Use \(\chi^2(3)\), 7.815 is the 5% upper critical value, so answer is 0.95.]
HLQ: Exam-Style Question with Worked Solution
Let \(X_1, X_2, X_3, X_4\) be independent random variables with \(X_i \sim N(0, i)\) for \(i = 1, 2, 3, 4\).
(a) Find the distribution of \(Y = \dfrac{X_1^2}{1} + \dfrac{X_2^2}{2} + \dfrac{X_3^2}{3}\). (3 marks)
(b) Find the distribution of \(W = \dfrac{X_1^2}{1} + \dfrac{X_2^2}{2} + \dfrac{X_3^2}{3} + \dfrac{X_4^2}{4}\). (3 marks)
(c) Let \(T = \dfrac{X_1}{\sqrt{W/4}}\). What is the distribution of \(T\)? (4 marks)
(d) Compute \(E[W]\) and \(\text{Var}(W)\). (4 marks)
Then \(Z_i^2 \sim \chi^2(1)\) and the \(Z_i\) are independent.
So \(Y = Z_1^2 + Z_2^2 + Z_3^2 \sim \chi^2(3)\).
\(W = Z_1^2 + Z_2^2 + Z_3^2 + Z_4^2 \sim \chi^2(4)\).
Here \(Z_1 \sim N(0,1)\) and \(W \sim \chi^2(4)\) are independent (\(W\) contains \(Z_1^2\) as part of the sum, but the t-distribution result still applies because the numerator variable \(Z_1\) is part of the denominator's chi-squared sum — however, this requires caution).
Actually, we need independence between numerator and denominator. Here \(Z_1\) appears in both the numerator and in \(W\), so they are not independent. This means \(T\) does not follow a standard t-distribution.
Alternative exam-sensible interpretation: If the question intended the numerator to be independent, we would use a different component. For instance, \(T = \dfrac{X_1}{\sqrt{(X_2^2/2 + X_3^2/3 + X_4^2/4)/3}}\) would be \(t(3)\) because \(Z_1\) is not in the denominator sum.
\(E[W] = 4\).
\(\text{Var}(W) = 2 \times 4 = 8\).
Alternatively, from linearity:
\(E[Z_i^2] = 1\) for each \(i\), so \(E[W] = 1 + 1 + 1 + 1 = 4\).
\(\text{Var}(Z_i^2) = 2\) for each \(i\), so \(\text{Var}(W) = 2 + 2 + 2 + 2 = 8\).