Menu
Advertisement
Advertisement
Advertisement

Estimation and Hypothesis Testing

ESTIMATION & INFERENCE

Chapter 9, Estimation and Hypothesis Testing , marks the bridge from theoretical probability and the study of sample statistics (Chapters 7 and 8) to the core application of statistical inference. This chapter explores methods used to leverage observed data to make formal statements, educated guesses, or concrete judgments about unknown characteristics of the underlying population distribution.


9.1 The Core Problem: Estimating Unknown Parameters

In statistics, we often assume an observed sample (X1,X2,,XnX_1, X_2, \dots, X_n) comes from a distribution with an unknown shape or parameter(s) (p1,p2,,pdp_1, p_2, \dots, p_d, denoted as pp). The goal of estimation is to use the sample data to provide a best guess for these unknown parameters.

Concept: Point Estimator

💡

Point Estimator

A point estimator is a function, g(X1,X2,,Xn)g(X_1, X_2, \dots, X_n), that takes the sample values as input and produces a single value (the estimate) intended to approximate the true parameter(s).

Example: If XX has an unknown mean μ\mu, the sample mean (Xˉ\bar{X}) is an unbiased estimator of μ\mu.


9.2 Method of Moments

The Method of Moments relies on matching sample moments to population moments.

Concept

  1. Sample Moment: mk=1ni=1nXikm_k = \frac{1}{n} \sum_{i=1}^{n} X_i^k
  2. Population Moment: μk=E[Xk]\mu_k = E[X^k] (function of unknown parameters).
  3. Solve: Set mk=μkm_k = \mu_k for k=1dk=1 \dots d and solve for parameters.

Example: Normal Distribution (μ,σ2\mu, \sigma^2)

  • μ^=Xˉ\hat{\mu} = \bar{X}
  • σ^2=1nXi2(Xˉ)2\hat{\sigma}^2 = \frac{1}{n} \sum X_i^2 - (\bar{X})^2

9.3 Maximum Likelihood Estimate (MLE)

The MLE method determines the parameter (p^\hat{p}) that makes the observed data “most likely”.

Concept: Likelihood Function

💡

Likelihood Function

The likelihood function, L(p)L(p), is the joint probability of the sample as a function of the parameters:

L(p;X1,,Xn)=i=1nf(Xip)L(p; X_1, \dots, X_n) = \prod_{i=1}^{n} f(X_i | p)

The MLE (p^\hat{p}) comes from maximising this function (or usually lnL\ln L).

Example: For Bernoulli(pp), the MLE is the sample proportion p^=Xˉ\hat{p} = \bar{X}.


9.4 Confidence Intervals

A confidence interval provides a range of values within which the true unknown parameter is expected to lie with a specified confidence level (β\beta).

Construction

  1. Known σ\sigma (Z-interval): Uses the Standard Normal distribution (ZZ). Z=n(Xˉμ)σZ = \frac{\sqrt{n}(\bar{X} - \mu)}{\sigma}

  2. Unknown σ\sigma (T-interval): Uses the t-distribution (tn1t_{n-1}) and sample deviation SS. T=n(Xˉμ)ST = \frac{\sqrt{n}(\bar{X} - \mu)}{S}

Key Difference: T-intervals are typically wider than Z-intervals, reflecting the extra uncertainty of estimating σ\sigma.


9.5 Hypothesis Testing

Hypothesis testing is a procedure to judge the plausibility of a conjecture about a parameter.

Core Concepts

  • Null Hypothesis (H0H_0): Baseline statement (e.g., μ=10\mu = 10).
  • Alternate Hypothesis (HaH_a): Contradiction (e.g., μ10\mu \neq 10).
  • Significance Level (α\alpha): Threshold for rejection (e.g., 0.05).
  • P-value: Probability of seeing data this extreme if H0H_0 were true.

Standardized Tests

Test NamePurposeTest StatisticNull Hypothesis
Z-testTest μ\mu (σ\sigma known/nn large)Z=Xˉμ0σ/nZ = \frac{\bar{X} - \mu_0}{\sigma/\sqrt{n}}μ=μ0\mu = \mu_0
T-testTest μ\mu (σ\sigma unknown)T=Xˉμ0S/nT = \frac{\bar{X} - \mu_0}{S/\sqrt{n}}μ=μ0\mu = \mu_0
χ2\chi^2-testTest σ2\sigma^2W=(n1)S2σ02W = \frac{(n-1)S^2}{\sigma_0^2}σ=σ0\sigma = \sigma_0
F-testCompare two variancesR=S12S22R = \frac{S_1^2}{S_2^2}σ1=σ2\sigma_1 = \sigma_2

Example: Z-test on Sample Mean

Q1

Hypothesis Test

Scenario: Normal population, σ=3.0\sigma=3.0, n=16n=16, Xˉ=10.2\bar{X}=10.2. Test H0:μ=9.5H_0: \mu = 9.5 vs Ha:μ>9.5H_a: \mu > 9.5 at α=0.05\alpha=0.05.

View Detailed Solution
  1. Test Statistic: Zobs=16(10.29.5)3.00.933Z_{\text{obs}} = \frac{\sqrt{16}(10.2 - 9.5)}{3.0} \approx 0.933
  2. P-value: P(Z0.933)0.175P(Z \geq 0.933) \approx 0.175.
  3. Conclusion: Since 0.175>0.050.175 > 0.05, we do not reject H0H_0.

Goodness of Fit (χ2\chi^2)

Test if observed categorical data matches expected counts.

💡

Chi-Square Statistic

χ2=j=1k(Yjnpj)2npj\chi^2 = \sum_{j=1}^{k} \frac{(Y_j - np_j)^2}{np_j}

Analogy: Estimation is like trying to guess the location of a hidden treasure. Point estimation gives you a single set of coordinates, while a confidence interval gives you a small map. Hypothesis testing is like testing a rumour about the treasure’s location (H0H_0) against the evidence found in the field.

Advertisement

All Chapters in this Book

Lesson 1

Basic Concepts

Foundational mathematical framework for probability, including definitions, axioms, conditional probability, and Bayes' Theorem.

Lesson 2

Sampling and Repeated Trials

Models based on repeated independent trials, focusing on Bernoulli trials and sampling methods.

Lesson 3

Discrete Random Variables

Formalizing random variables, probability mass functions, and independence.

Lesson 4

Summarizing Discrete Random Variables

Deriving numerical characteristics—expected value, variance, and standard deviation—to summarize behavior of discrete random variables.

Lesson 5

Continuous Probabilities and Random Variables

Transitioning from discrete sums to continuous integrals, density functions, and key distributions like Normal and Exponential.

Lesson 6

Summarising Continuous Random Variables

Extending expected value and variance to continuous variables, exploring Moment Generating Functions and Bivariate Normal distributions.

Lesson 7

Sampling and Descriptive Statistics

Transitioning from probability to statistics: using sample data to estimate population parameters like mean and variance.

Lesson 8

Sampling Distributions and Limit Theorems

The theoretical foundations of inference: Joint Distributions, Weak Law of Large Numbers (WLLN), and geometrical convergence via the Central Limit Theorem (CLT).

Lesson 9

Estimation and Hypothesis Testing

The core of statistical inference: Method of Moments, Maximum Likelihood, Confidence Intervals, and Hypothesis Testing.

Lesson 10

Linear Regression

Modeling linear relationships, least squares, and regression inference.

Sponsored Content

finding (solutions) x

A public notebook and learning hub.