Menu

Discrete Random Variables

Discrete Random Variables

Chapter 3 marks a critical shift in perspective from viewing outcomes as abstract events in a sample space to seeing them as numerical values generated by functions called “random variables” . This conceptual framework allows the use of mathematical analysis to describe and summarize the behavior of probabilistic experiments.

The chapter is structured around defining random variables, understanding how they transmit probability, classifying common distribution types, and examining the relationships between multiple variables.


3.1 Random Variables: Functions of Chance

In Chapter 1, we defined an experiment resulting in an outcome from a sample space SS. Often, we are interested not in the raw outcome (e.g., {Heads, Tails}) but in a numerical feature derived from it.

Concept Explanation: Random Variables

💡

Random Variable (X)

A Random Variable (XX) is simply a function that maps every outcome in the sample space SS to a real number.

A variable is specifically classified as discrete if its output values (its range, TT) form a countable (or finite) subset of the real numbers.

ExperimentSample Space (SS)Random Variable (XX)Range (TT)
Flip coin 3 times{hhh,hht,,ttt}\{hhh, hht, \dots, ttt\}Total number of heads{0,1,2,3}\{0, 1, 2, 3\}
Roll pair of dice36 pairs (e.g., (1, 1))Larger of the two values{1,2,3,4,5,6}\{1, 2, 3, 4, 5, 6\}

Concept: Probability Mass Function (PMF)

For a discrete random variable XX, its distribution is completely defined by knowing the probability associated with each possible value tt in its range TT.

PMF Definition


fX(t)=P(X=t)f_X(t) = P(X = t)


for all tTt \in T

How it works

The PMF assigns a probability “mass” to each discrete value.

Sum of all masses = 1.

P(XA)=tAfX(t)P(X \in A) = \sum_{t \in A} f_X(t)

Example Question and Solution (PMF)

Q1

Coin Flip PMF

Question: If a coin is flipped three times, and XX is the total number of heads, find the PMF of XX.

📝 View Detailed Solution

Solution: The sample space SS has S=8|S|=8 equally likely outcomes.

  • P(X=0)=P({ttt})=1/8P(X=0) = P(\{ttt\}) = 1/8.
  • P(X=1)=P({htt,tht,tth})=3/8P(X=1) = P(\{htt, tht, tth\}) = 3/8.
  • P(X=2)=P({hht,hth,thh})=3/8P(X=2) = P(\{hht, hth, thh\}) = 3/8.
  • P(X=3)=P({hhh})=1/8P(X=3) = P(\{hhh\}) = 1/8.

The PMF is fX(0)=1/8,fX(1)=3/8,fX(2)=3/8,fX(3)=1/8f_X(0)=1/8, f_X(1)=3/8, f_X(2)=3/8, f_X(3)=1/8.

Common Discrete Distributions

DistributionNotationPMF, P(X=k)P(X=k)Context
BernoulliXBernoulli(p)X \sim Bernoulli(p)pp (if k=1k=1)Single trial
BinomialXBinomial(n,p)X \sim Binomial(n, p)(nk)pk(1p)nk\binom{n}{k} p^k (1-p)^{n-k}kk successes in nn trials
GeometricXGeometric(p)X \sim Geometric(p)p(1p)k1p (1-p)^{k-1}Trials until first success
PoissonXPoisson(λ)X \sim Poisson(\lambda)eλλkk!\frac{e^{-\lambda} \lambda^k}{k!}Rare events (nn large, pp small)

3.2 Relationships Between Random Variables

When multiple random variables are defined on the same sample space, we analyze how they interact.

Concept: Independence

💡

Independence

Two random variables, XX and YY, are independent if the occurrence of any event related to XX does not affect probabilities related to YY.

P(X=t,Y=u)=P(X=t)P(Y=u)\mathbf{P(X=t, Y=u) = P(X=t) P(Y=u)}

A sequence is i.i.d. (“independent and identically distributed”) if all variables are mutually independent and share the exact same distribution.

Joint & Conditional Distributions

The joint distribution captures the relationship between variables: Q((a,b))=P(X=a,Y=b)Q((a, b)) = P(X=a, Y=b).

Example Question and Solution (Joint Table)

Q2

Joint Probability Table

Let XX and YY be defined by the following joint distribution table. Find P(X=1Y=0)P(X=1 | Y=0).

X=1X=2Total P(Y=b)
Y=01/41/83/8
Y=11/41/44/8
Y=201/81/8
Total1/21/21
📝 View Detailed Solution

Solution: Using the formula for conditional probability: P(X=1Y=0)=P(X=1,Y=0)P(Y=0)=1/43/8P(X=1 | Y=0) = \frac{P(X=1, Y=0)}{P(Y=0)} = \frac{1/4}{3/8}

P(X=1Y=0)=14×83=23P(X=1 | Y=0) = \frac{1}{4} \times \frac{8}{3} = \frac{2}{3}

Key Property: Memoryless Distribution

M

Memoryless


Past failures don’t
affect future success.

M
Geometric
💡

Memoryless Property

The Geometric distribution is memoryless. If XX is the number of trials until the first success:

P(X>n+mX>n)=P(X>m)\mathbf{P(X > n+m | X > n) = P(X > m)}

The sequence “starts over” every time a failure occurs.


3.3 Functions of Random Variables

A function of a random variable, Z=f(X)Z = f(X), is itself a new random variable.

Convolution (Sum of Independent Variables)

When examining the sum of two independent random variables, Z=X+YZ = X+Y, the distribution P(Z=n)P(Z=n) is calculated using convolution.

Convolution Sum


P(Z=n)=j=0nP(X=j)P(Y=nj)P(Z=n) = \sum_{j=0}^{n} P(X=j) \cdot P(Y=n-j)

Logic

We sum the probabilities of all possible ways to get a total of nn:

X=0,Y=nX=0, Y=n

X=1,Y=n1X=1, Y=n-1

…etc…

Example Application (Sum of Poissons)

Q3

Sum of Poissons

Question: If XPoisson(λ1)X \sim Poisson(\lambda_1) and YPoisson(λ2)Y \sim Poisson(\lambda_2) are independent, what is the distribution of their sum Z=X+YZ = X+Y?

📝 View Detailed Solution

Using the convolution sum and algebraic manipulation (including the binomial expansion), it can be proven that the sum of two independent Poisson variables is also a Poisson variable:

ZPoisson(λ1+λ2)\mathbf{Z \sim Poisson(\lambda_1 + \lambda_2)}