Estimation and Hypothesis Testing
ESTIMATION & INFERENCE
Chapter 9, Estimation and Hypothesis Testing , marks the bridge from theoretical probability and the study of sample statistics (Chapters 7 and 8) to the core application of statistical inference. This chapter explores methods used to leverage observed data to make formal statements, educated guesses, or concrete judgments about unknown characteristics of the underlying population distribution.
9.1 The Core Problem: Estimating Unknown Parameters
In statistics, we often assume an observed sample () comes from a distribution with an unknown shape or parameter(s) (, denoted as ). The goal of estimation is to use the sample data to provide a best guess for these unknown parameters.
Concept: Point Estimator
Point Estimator
A point estimator is a function, , that takes the sample values as input and produces a single value (the estimate) intended to approximate the true parameter(s).
Example: If has an unknown mean , the sample mean () is an unbiased estimator of .
9.2 Method of Moments
The Method of Moments relies on matching sample moments to population moments.
Concept
- Sample Moment:
- Population Moment: (function of unknown parameters).
- Solve: Set for and solve for parameters.
Example: Normal Distribution ()
9.3 Maximum Likelihood Estimate (MLE)
The MLE method determines the parameter () that makes the observed data “most likely”.
Concept: Likelihood Function
Likelihood Function
The likelihood function, , is the joint probability of the sample as a function of the parameters:
The MLE () comes from maximising this function (or usually ).
Example: For Bernoulli(), the MLE is the sample proportion .
9.4 Confidence Intervals
A confidence interval provides a range of values within which the true unknown parameter is expected to lie with a specified confidence level ().
Construction
-
Known (Z-interval): Uses the Standard Normal distribution ().
-
Unknown (T-interval): Uses the t-distribution () and sample deviation .
Key Difference: T-intervals are typically wider than Z-intervals, reflecting the extra uncertainty of estimating .
9.5 Hypothesis Testing
Hypothesis testing is a procedure to judge the plausibility of a conjecture about a parameter.
Core Concepts
- Null Hypothesis (): Baseline statement (e.g., ).
- Alternate Hypothesis (): Contradiction (e.g., ).
- Significance Level (): Threshold for rejection (e.g., 0.05).
- P-value: Probability of seeing data this extreme if were true.
Standardized Tests
| Test Name | Purpose | Test Statistic | Null Hypothesis |
|---|---|---|---|
| Z-test | Test ( known/ large) | ||
| T-test | Test ( unknown) | ||
| -test | Test | ||
| F-test | Compare two variances |
Example: Z-test on Sample Mean
Hypothesis Test
Scenario: Normal population, , , . Test vs at .
📝 View Detailed Solution ▼
- Test Statistic:
- P-value: .
- Conclusion: Since , we do not reject .
Goodness of Fit ()
Test if observed categorical data matches expected counts.
Chi-Square Statistic
Analogy: Estimation is like trying to guess the location of a hidden treasure. Point estimation gives you a single set of coordinates, while a confidence interval gives you a small map. Hypothesis testing is like testing a rumour about the treasure’s location () against the evidence found in the field.