Menu

Estimation and Hypothesis Testing

ESTIMATION & INFERENCE

Chapter 9, Estimation and Hypothesis Testing , marks the bridge from theoretical probability and the study of sample statistics (Chapters 7 and 8) to the core application of statistical inference. This chapter explores methods used to leverage observed data to make formal statements, educated guesses, or concrete judgments about unknown characteristics of the underlying population distribution.


9.1 The Core Problem: Estimating Unknown Parameters

In statistics, we often assume an observed sample (X1,X2,,XnX_1, X_2, \dots, X_n) comes from a distribution with an unknown shape or parameter(s) (p1,p2,,pdp_1, p_2, \dots, p_d, denoted as pp). The goal of estimation is to use the sample data to provide a best guess for these unknown parameters.

Concept: Point Estimator

💡

Point Estimator

A point estimator is a function, g(X1,X2,,Xn)g(X_1, X_2, \dots, X_n), that takes the sample values as input and produces a single value (the estimate) intended to approximate the true parameter(s).

Example: If XX has an unknown mean μ\mu, the sample mean (Xˉ\bar{X}) is an unbiased estimator of μ\mu.


9.2 Method of Moments

The Method of Moments relies on matching sample moments to population moments.

Concept

  1. Sample Moment: mk=1ni=1nXikm_k = \frac{1}{n} \sum_{i=1}^{n} X_i^k
  2. Population Moment: μk=E[Xk]\mu_k = E[X^k] (function of unknown parameters).
  3. Solve: Set mk=μkm_k = \mu_k for k=1dk=1 \dots d and solve for parameters.

Example: Normal Distribution (μ,σ2\mu, \sigma^2)

  • μ^=Xˉ\hat{\mu} = \bar{X}
  • σ^2=1nXi2(Xˉ)2\hat{\sigma}^2 = \frac{1}{n} \sum X_i^2 - (\bar{X})^2

9.3 Maximum Likelihood Estimate (MLE)

The MLE method determines the parameter (p^\hat{p}) that makes the observed data “most likely”.

Concept: Likelihood Function

💡

Likelihood Function

The likelihood function, L(p)L(p), is the joint probability of the sample as a function of the parameters:

L(p;X1,,Xn)=i=1nf(Xip)L(p; X_1, \dots, X_n) = \prod_{i=1}^{n} f(X_i | p)

The MLE (p^\hat{p}) comes from maximising this function (or usually lnL\ln L).

Example: For Bernoulli(pp), the MLE is the sample proportion p^=Xˉ\hat{p} = \bar{X}.


9.4 Confidence Intervals

A confidence interval provides a range of values within which the true unknown parameter is expected to lie with a specified confidence level (β\beta).

Construction

  1. Known σ\sigma (Z-interval): Uses the Standard Normal distribution (ZZ). Z=n(Xˉμ)σZ = \frac{\sqrt{n}(\bar{X} - \mu)}{\sigma}

  2. Unknown σ\sigma (T-interval): Uses the t-distribution (tn1t_{n-1}) and sample deviation SS. T=n(Xˉμ)ST = \frac{\sqrt{n}(\bar{X} - \mu)}{S}

Key Difference: T-intervals are typically wider than Z-intervals, reflecting the extra uncertainty of estimating σ\sigma.


9.5 Hypothesis Testing

Hypothesis testing is a procedure to judge the plausibility of a conjecture about a parameter.

Core Concepts

  • Null Hypothesis (H0H_0): Baseline statement (e.g., μ=10\mu = 10).
  • Alternate Hypothesis (HaH_a): Contradiction (e.g., μ10\mu \neq 10).
  • Significance Level (α\alpha): Threshold for rejection (e.g., 0.05).
  • P-value: Probability of seeing data this extreme if H0H_0 were true.

Standardized Tests

Test NamePurposeTest StatisticNull Hypothesis
Z-testTest μ\mu (σ\sigma known/nn large)Z=Xˉμ0σ/nZ = \frac{\bar{X} - \mu_0}{\sigma/\sqrt{n}}μ=μ0\mu = \mu_0
T-testTest μ\mu (σ\sigma unknown)T=Xˉμ0S/nT = \frac{\bar{X} - \mu_0}{S/\sqrt{n}}μ=μ0\mu = \mu_0
χ2\chi^2-testTest σ2\sigma^2W=(n1)S2σ02W = \frac{(n-1)S^2}{\sigma_0^2}σ=σ0\sigma = \sigma_0
F-testCompare two variancesR=S12S22R = \frac{S_1^2}{S_2^2}σ1=σ2\sigma_1 = \sigma_2

Example: Z-test on Sample Mean

Q1

Hypothesis Test

Scenario: Normal population, σ=3.0\sigma=3.0, n=16n=16, Xˉ=10.2\bar{X}=10.2. Test H0:μ=9.5H_0: \mu = 9.5 vs Ha:μ>9.5H_a: \mu > 9.5 at α=0.05\alpha=0.05.

📝 View Detailed Solution
  1. Test Statistic: Zobs=16(10.29.5)3.00.933Z_{\text{obs}} = \frac{\sqrt{16}(10.2 - 9.5)}{3.0} \approx 0.933
  2. P-value: P(Z0.933)0.175P(Z \geq 0.933) \approx 0.175.
  3. Conclusion: Since 0.175>0.050.175 > 0.05, we do not reject H0H_0.

Goodness of Fit (χ2\chi^2)

Test if observed categorical data matches expected counts.

💡

Chi-Square Statistic

χ2=j=1k(Yjnpj)2npj\chi^2 = \sum_{j=1}^{k} \frac{(Y_j - np_j)^2}{np_j}

Analogy: Estimation is like trying to guess the location of a hidden treasure. Point estimation gives you a single set of coordinates, while a confidence interval gives you a small map. Hypothesis testing is like testing a rumour about the treasure’s location (H0H_0) against the evidence found in the field.