Basic Concepts
B A S I C C O N C E P T S
Chapter 1 of the “Stats 2 Book” establishes the fundamental vocabulary and axiomatic framework necessary for studying probability and statistics. This chapter introduces the structure needed to discuss the likelihood of occurrences.
Here is a detailed explanation of the key concepts, followed by illustrative examples and exercises based on the source material:
1.1 Definitions and Properties
The foundation of probability theory rests on defining what is possible before determining what is likely.
Key Definitions
| Concept | Definition | Example |
|---|---|---|
| Sample Space () | A set listing all possibilities (outcomes) that might occur. | When rolling a six-sided die, . |
| Outcome | An element of the sample space . | The result of rolling a die, e.g., ‘4’. |
| Experiment | The process of actually selecting one of the outcomes listed. | Flipping a coin or waiting for the winner of the World Cup. |
| Event () | Any subset of the sample space . | Rolling a number > 2: . |
Probability Axioms
A probability () is a function that assigns a chance (a number between 0 and 1) to each event . This formally relies on Kolmogorov’s axioms.
Axiom 1: Certainty
The probability of the entire sample space is 1. Interpretation: There is a 100% chance that an experiment will result in some outcome included in .
Axiom 2: Additivity
For any countable collection of disjoint events (), their combined probability is the sum of their individual probabilities. Interpretation: If events cannot happen simultaneously, their probabilities add up.
Basic Properties
From the two fundamental axioms, several properties can be proven that simplify probability calculations:
| Property | Formula | Description |
|---|---|---|
| Empty Set | The probability of nothing happening is zero. | |
| Finite Additivity | Sum rule for finite disjoint events. | |
| Monotonicity | If | Subsets cannot be more likely than supersets. |
| Difference Rule | Prob. of occurring but not (if ). | |
| Complement Rule | Prob. that does NOT occur. | |
| General Addition | Subtract intersection to avoid double counting. |
Example and Solution
Coin Flip Axioms
A fair coin flip has a sample space . Use the axioms to show that the probability of observing heads is 0.5.
View Detailed Solution ▼
- Let and be two disjoint events.
- Since the coin is “fair,” for some value .
- The union is the sample space: .
- Using Axiom 1 and Axiom 2:
- Substituting : .
Fishing Tonnage
A town’s fishing fleet has a 35% chance of catching over 400 tons () and a 10% chance of catching over 500 tons (). How likely is it that they will catch between 400 and 500 tons ?
View Detailed Solution ▼
- Variable (“over 500”) is a subset of (“over 400”).
- “Between 400 and 500” is the difference set .
- Using the Difference Rule:
Answer: There is a 25% chance that between 400 and 500 tons of fish will be caught.
1.2 Equally Likely Outcomes
When outcomes are finite and equally likely, we have a Uniform Distribution.
Running probability calculations in this setting becomes a pure counting problem.
Rolling Two Dice
Two dice are rolled. How likely is it that their sum will equal eight?
View Detailed Solution ▼
- Total sample space .
- Event (sum is 8): .
- Count .
- Probability:
Group Selection
A group of 12 people includes Grant and Dilip. Pick 3 randomly. How likely is it to include Grant, but not Dilip?
Combinatorial Solution ▼
- Total Outcomes: Choose 3 from 12. .
- Event : Grant is fixed (1 way). Dilip is out. We need 2 more from the remaining 10.
- Count : Choose 2 from 10. .
- Probability:
1.3 Conditional Probability
How is likelihood “altered” by knowledge that another event has occurred?
Conditional Probability Formula
Provided
Bayes’ Theorem
One of the most powerful theorems in statistics allows us to reverse conditional probabilities.
Bayes' Theorem
Updates belief given new evidence .
The Swine Flu Test
- Detects Flu 95% of the time if infected: .
- False Positive rate is 2%: .
- Population rate is 1%: .
If a person tests positive, what is the probability they actually have the flu?
Bayesian Check ▼
- Let , .
- We want .
- Apply Bayes’ Theorem:
Result: Only a 32.4% chance they actually have the flu!
1.4 Independence
Events are independent if the occurrence of one has no effect on the other.
1.5 Using R
R is essential for complex calculations.
| Function | Syntax | Purpose |
|---|---|---|
| Vector Creation | c(1, 2, 3) | Creates a list of numbers. |
| Sequence | 1:100 | Creates integers from 1 to 100. |
| Combinations | choose(n, k) | Calculates . |
| Summation | sum(x) | Adds all elements in vector . |
All Chapters in this Book
Basic Concepts
Foundational mathematical framework for probability, including definitions, axioms, conditional probability, and Bayes' Theorem.
Sampling and Repeated Trials
Models based on repeated independent trials, focusing on Bernoulli trials and sampling methods.
Discrete Random Variables
Formalizing random variables, probability mass functions, and independence.
Summarizing Discrete Random Variables
Deriving numerical characteristics—expected value, variance, and standard deviation—to summarize behavior of discrete random variables.
Continuous Probabilities and Random Variables
Transitioning from discrete sums to continuous integrals, density functions, and key distributions like Normal and Exponential.
Summarising Continuous Random Variables
Extending expected value and variance to continuous variables, exploring Moment Generating Functions and Bivariate Normal distributions.
Sampling and Descriptive Statistics
Transitioning from probability to statistics: using sample data to estimate population parameters like mean and variance.
Sampling Distributions and Limit Theorems
The theoretical foundations of inference: Joint Distributions, Weak Law of Large Numbers (WLLN), and geometrical convergence via the Central Limit Theorem (CLT).
Estimation and Hypothesis Testing
The core of statistical inference: Method of Moments, Maximum Likelihood, Confidence Intervals, and Hypothesis Testing.
Linear Regression
Modeling linear relationships, least squares, and regression inference.