Discrete Random Variables
Discrete Random Variables
Chapter 3 marks a critical shift in perspective from viewing outcomes as abstract events in a sample space to seeing them as numerical values generated by functions called “random variables” . This conceptual framework allows the use of mathematical analysis to describe and summarize the behavior of probabilistic experiments.
The chapter is structured around defining random variables, understanding how they transmit probability, classifying common distribution types, and examining the relationships between multiple variables.
3.1 Random Variables: Functions of Chance
In Chapter 1, we defined an experiment resulting in an outcome from a sample space . Often, we are interested not in the raw outcome (e.g., {Heads, Tails}) but in a numerical feature derived from it.
Concept Explanation: Random Variables
Random Variable (X)
A Random Variable () is simply a function that maps every outcome in the sample space to a real number.
A variable is specifically classified as discrete if its output values (its range, ) form a countable (or finite) subset of the real numbers.
| Experiment | Sample Space () | Random Variable () | Range () |
|---|---|---|---|
| Flip coin 3 times | Total number of heads | ||
| Roll pair of dice | 36 pairs (e.g., (1, 1)) | Larger of the two values |
Concept: Probability Mass Function (PMF)
For a discrete random variable , its distribution is completely defined by knowing the probability associated with each possible value in its range .
PMF Definition
for all
How it works
The PMF assigns a probability “mass” to each discrete value.
Sum of all masses = 1.
Example Question and Solution (PMF)
Coin Flip PMF
Question: If a coin is flipped three times, and is the total number of heads, find the PMF of .
View Detailed Solution ▼
Solution: The sample space has equally likely outcomes.
- .
- .
- .
- .
The PMF is .
Common Discrete Distributions
| Distribution | Notation | PMF, | Context |
|---|---|---|---|
| Bernoulli | (if ) | Single trial | |
| Binomial | successes in trials | ||
| Geometric | Trials until first success | ||
| Poisson | Rare events ( large, small) |
3.2 Relationships Between Random Variables
When multiple random variables are defined on the same sample space, we analyze how they interact.
Concept: Independence
Independence
Two random variables, and , are independent if the occurrence of any event related to does not affect probabilities related to .
A sequence is i.i.d. (“independent and identically distributed”) if all variables are mutually independent and share the exact same distribution.
Joint & Conditional Distributions
The joint distribution captures the relationship between variables: .
Example Question and Solution (Joint Table)
Joint Probability Table
Let and be defined by the following joint distribution table. Find .
| X=1 | X=2 | Total P(Y=b) | |
|---|---|---|---|
| Y=0 | 1/4 | 1/8 | 3/8 |
| Y=1 | 1/4 | 1/4 | 4/8 |
| Y=2 | 0 | 1/8 | 1/8 |
| Total | 1/2 | 1/2 | 1 |
View Detailed Solution ▼
Solution: Using the formula for conditional probability:
Key Property: Memoryless Distribution
Memoryless Property
The Geometric distribution is memoryless. If is the number of trials until the first success:
The sequence “starts over” every time a failure occurs.
3.3 Functions of Random Variables
A function of a random variable, , is itself a new random variable.
Convolution (Sum of Independent Variables)
When examining the sum of two independent random variables, , the distribution is calculated using convolution.
Convolution Sum
Logic
We sum the probabilities of all possible ways to get a total of :
…etc…
Example Application (Sum of Poissons)
Sum of Poissons
Question: If and are independent, what is the distribution of their sum ?
View Detailed Solution ▼
Using the convolution sum and algebraic manipulation (including the binomial expansion), it can be proven that the sum of two independent Poisson variables is also a Poisson variable:
All Chapters in this Book
Basic Concepts
Foundational mathematical framework for probability, including definitions, axioms, conditional probability, and Bayes' Theorem.
Sampling and Repeated Trials
Models based on repeated independent trials, focusing on Bernoulli trials and sampling methods.
Discrete Random Variables
Formalizing random variables, probability mass functions, and independence.
Summarizing Discrete Random Variables
Deriving numerical characteristics—expected value, variance, and standard deviation—to summarize behavior of discrete random variables.
Continuous Probabilities and Random Variables
Transitioning from discrete sums to continuous integrals, density functions, and key distributions like Normal and Exponential.
Summarising Continuous Random Variables
Extending expected value and variance to continuous variables, exploring Moment Generating Functions and Bivariate Normal distributions.
Sampling and Descriptive Statistics
Transitioning from probability to statistics: using sample data to estimate population parameters like mean and variance.
Sampling Distributions and Limit Theorems
The theoretical foundations of inference: Joint Distributions, Weak Law of Large Numbers (WLLN), and geometrical convergence via the Central Limit Theorem (CLT).
Estimation and Hypothesis Testing
The core of statistical inference: Method of Moments, Maximum Likelihood, Confidence Intervals, and Hypothesis Testing.
Linear Regression
Modeling linear relationships, least squares, and regression inference.