Discrete Random Variables
Discrete Random Variables
Chapter 3 marks a critical shift in perspective from viewing outcomes as abstract events in a sample space to seeing them as numerical values generated by functions called “random variables” . This conceptual framework allows the use of mathematical analysis to describe and summarize the behavior of probabilistic experiments.
The chapter is structured around defining random variables, understanding how they transmit probability, classifying common distribution types, and examining the relationships between multiple variables.
3.1 Random Variables: Functions of Chance
In Chapter 1, we defined an experiment resulting in an outcome from a sample space . Often, we are interested not in the raw outcome (e.g., {Heads, Tails}) but in a numerical feature derived from it.
Concept Explanation: Random Variables
Random Variable (X)
A Random Variable () is simply a function that maps every outcome in the sample space to a real number.
A variable is specifically classified as discrete if its output values (its range, ) form a countable (or finite) subset of the real numbers.
| Experiment | Sample Space () | Random Variable () | Range () |
|---|---|---|---|
| Flip coin 3 times | Total number of heads | ||
| Roll pair of dice | 36 pairs (e.g., (1, 1)) | Larger of the two values |
Concept: Probability Mass Function (PMF)
For a discrete random variable , its distribution is completely defined by knowing the probability associated with each possible value in its range .
PMF Definition
for all
How it works
The PMF assigns a probability “mass” to each discrete value.
Sum of all masses = 1.
Example Question and Solution (PMF)
Coin Flip PMF
Question: If a coin is flipped three times, and is the total number of heads, find the PMF of .
📝 View Detailed Solution ▼
Solution: The sample space has equally likely outcomes.
- .
- .
- .
- .
The PMF is .
Common Discrete Distributions
| Distribution | Notation | PMF, | Context |
|---|---|---|---|
| Bernoulli | (if ) | Single trial | |
| Binomial | successes in trials | ||
| Geometric | Trials until first success | ||
| Poisson | Rare events ( large, small) |
3.2 Relationships Between Random Variables
When multiple random variables are defined on the same sample space, we analyze how they interact.
Concept: Independence
Independence
Two random variables, and , are independent if the occurrence of any event related to does not affect probabilities related to .
A sequence is i.i.d. (“independent and identically distributed”) if all variables are mutually independent and share the exact same distribution.
Joint & Conditional Distributions
The joint distribution captures the relationship between variables: .
Example Question and Solution (Joint Table)
Joint Probability Table
Let and be defined by the following joint distribution table. Find .
| X=1 | X=2 | Total P(Y=b) | |
|---|---|---|---|
| Y=0 | 1/4 | 1/8 | 3/8 |
| Y=1 | 1/4 | 1/4 | 4/8 |
| Y=2 | 0 | 1/8 | 1/8 |
| Total | 1/2 | 1/2 | 1 |
📝 View Detailed Solution ▼
Solution: Using the formula for conditional probability:
Key Property: Memoryless Distribution
Memoryless Property
The Geometric distribution is memoryless. If is the number of trials until the first success:
The sequence “starts over” every time a failure occurs.
3.3 Functions of Random Variables
A function of a random variable, , is itself a new random variable.
Convolution (Sum of Independent Variables)
When examining the sum of two independent random variables, , the distribution is calculated using convolution.
Convolution Sum
Logic
We sum the probabilities of all possible ways to get a total of :
…etc…
Example Application (Sum of Poissons)
Sum of Poissons
Question: If and are independent, what is the distribution of their sum ?
📝 View Detailed Solution ▼
Using the convolution sum and algebraic manipulation (including the binomial expansion), it can be proven that the sum of two independent Poisson variables is also a Poisson variable: