Estimation and Hypothesis Testing
ESTIMATION & INFERENCE
Chapter 9, Estimation and Hypothesis Testing , marks the bridge from theoretical probability and the study of sample statistics (Chapters 7 and 8) to the core application of statistical inference. This chapter explores methods used to leverage observed data to make formal statements, educated guesses, or concrete judgments about unknown characteristics of the underlying population distribution.
9.1 The Core Problem: Estimating Unknown Parameters
In statistics, we often assume an observed sample () comes from a distribution with an unknown shape or parameter(s) (, denoted as ). The goal of estimation is to use the sample data to provide a best guess for these unknown parameters.
Concept: Point Estimator
Point Estimator
A point estimator is a function, , that takes the sample values as input and produces a single value (the estimate) intended to approximate the true parameter(s).
Example: If has an unknown mean , the sample mean () is an unbiased estimator of .
9.2 Method of Moments
The Method of Moments relies on matching sample moments to population moments.
Concept
- Sample Moment:
- Population Moment: (function of unknown parameters).
- Solve: Set for and solve for parameters.
Example: Normal Distribution ()
9.3 Maximum Likelihood Estimate (MLE)
The MLE method determines the parameter () that makes the observed data “most likely”.
Concept: Likelihood Function
Likelihood Function
The likelihood function, , is the joint probability of the sample as a function of the parameters:
The MLE () comes from maximising this function (or usually ).
Example: For Bernoulli(), the MLE is the sample proportion .
9.4 Confidence Intervals
A confidence interval provides a range of values within which the true unknown parameter is expected to lie with a specified confidence level ().
Construction
-
Known (Z-interval): Uses the Standard Normal distribution ().
-
Unknown (T-interval): Uses the t-distribution () and sample deviation .
Key Difference: T-intervals are typically wider than Z-intervals, reflecting the extra uncertainty of estimating .
9.5 Hypothesis Testing
Hypothesis testing is a procedure to judge the plausibility of a conjecture about a parameter.
Core Concepts
- Null Hypothesis (): Baseline statement (e.g., ).
- Alternate Hypothesis (): Contradiction (e.g., ).
- Significance Level (): Threshold for rejection (e.g., 0.05).
- P-value: Probability of seeing data this extreme if were true.
Standardized Tests
| Test Name | Purpose | Test Statistic | Null Hypothesis |
|---|---|---|---|
| Z-test | Test ( known/ large) | ||
| T-test | Test ( unknown) | ||
| -test | Test | ||
| F-test | Compare two variances |
Example: Z-test on Sample Mean
Hypothesis Test
Scenario: Normal population, , , . Test vs at .
View Detailed Solution ▼
- Test Statistic:
- P-value: .
- Conclusion: Since , we do not reject .
Goodness of Fit ()
Test if observed categorical data matches expected counts.
Chi-Square Statistic
Analogy: Estimation is like trying to guess the location of a hidden treasure. Point estimation gives you a single set of coordinates, while a confidence interval gives you a small map. Hypothesis testing is like testing a rumour about the treasure’s location () against the evidence found in the field.
All Chapters in this Book
Basic Concepts
Foundational mathematical framework for probability, including definitions, axioms, conditional probability, and Bayes' Theorem.
Sampling and Repeated Trials
Models based on repeated independent trials, focusing on Bernoulli trials and sampling methods.
Discrete Random Variables
Formalizing random variables, probability mass functions, and independence.
Summarizing Discrete Random Variables
Deriving numerical characteristics—expected value, variance, and standard deviation—to summarize behavior of discrete random variables.
Continuous Probabilities and Random Variables
Transitioning from discrete sums to continuous integrals, density functions, and key distributions like Normal and Exponential.
Summarising Continuous Random Variables
Extending expected value and variance to continuous variables, exploring Moment Generating Functions and Bivariate Normal distributions.
Sampling and Descriptive Statistics
Transitioning from probability to statistics: using sample data to estimate population parameters like mean and variance.
Sampling Distributions and Limit Theorems
The theoretical foundations of inference: Joint Distributions, Weak Law of Large Numbers (WLLN), and geometrical convergence via the Central Limit Theorem (CLT).
Estimation and Hypothesis Testing
The core of statistical inference: Method of Moments, Maximum Likelihood, Confidence Intervals, and Hypothesis Testing.
Linear Regression
Modeling linear relationships, least squares, and regression inference.