Menu

Continuous Probabilities and Random Variables

CONTINUOUS PROBABILITIES

Chapter 5 of your statistics textbook, Continuous Probabilities and Random Variables , marks a significant transition in how we define and calculate probabilities. Unlike discrete variables (like counts or die rolls) which have a countable number of outcomes, continuous variables (like temperature, time, or height) can take on any value within a range. This shift requires abandoning sums and introducing the concept of density and integration.


5.1 The Necessary Shift: From Sums to Integrals

In discrete probability, every outcome has a specific chance of occurring, and we calculate the probability of an event by summing up the likelihoods of the relevant outcomes. This approach fails entirely in the continuous world.

Concept: Zero Probability for Individual Outcomes

If you try to assign a uniform probability pp to every single outcome xx in an uncountable interval (like (0, 1)), the total probability P(S)P(S) would diverge to infinity if p>0p>0.

Therefore, for any continuous random variable XX, the probability of landing on any single exact point is zero: P(X=a)=0P(X = a) = 0.

Question: If P(X=x)=0P(X=x) = 0 for every point xx, how can the total probability P(S)P(S) still be 1?

Explanation: The fundamental axioms of probability only allow us to sum up the probabilities of a countable collection of disjoint events. Since the interval (0, 1) contains an uncountable collection of points, the probability of the entire space is still 1, even though the probability of any individual point is zero.

Concept: The Probability Density Function (PDF)

Since individual points have zero probability, we must define probability over intervals or sets, which is done using a function called the probability density function (f(x)f(x)), or simply the density.

  1. Non-Negative: The density must be non-negative for all values of xx (probabilities cannot be negative).
  2. Total Area is One: The total area under the density curve must equal one (representing 100% certainty that the outcome occurs somewhere on the real line): f(x)dx=1\int_{-\infty}^{\infty} f(x) dx = 1.
💡

Probability via Integration

The probability of a continuous random variable XX falling into an event AA is found by calculating the area under the density curve over that set AA:

P(XA)=Af(x)dxP(X \in A) = \int_A f(x) dx

Example: Non-Uniform Density

Q1

Area under Density

A density function f(x)f(x) may not be constant, meaning equal-length intervals might have different probabilities.

If a density is given by f(x)=3x2f(x) = 3x^2 for 0<x<10 < x < 1:

  • P([0.2,0.4])=0.20.43x2dx=0.056P([0.2, 0.4]) = \int_{0.2}^{0.4} 3x^2 dx = 0.056
  • P([0.6,0.8])=0.60.83x2dx=0.296P([0.6, 0.8]) = \int_{0.6}^{0.8} 3x^2 dx = 0.296

The second interval (which is further away from zero) has a much higher probability, reflecting that the density function is larger in that region.


5.2 & 5.3 Summarizing Continuous Variables

Concept: The Cumulative Distribution Function (CDF)

The distribution function or Cumulative Distribution Function (F(x)F(x)) of a random variable XX gives the probability that XX is less than or equal to a specific value xx:

F(x)=P(Xx)F(x) = P(X \leq x)

For a continuous random variable with density f(x)f(x), the CDF is the integral of the density up to point xx: F(x)=xf(x)dxF(x) = \int_{-\infty}^{x} f(x) dx

This relationship is crucial, as the CDF is differentiable wherever the density f(x)f(x) is continuous, and F(x)=f(x)F'(x) = f(x). This means that the PDF and the CDF contain the exact same information about the distribution of XX.

Key Continuous Distributions

DistributionNotationDensity Function (f(x)f(x))Core Application
UniformUniform(a,b)Uniform(a, b)1ba\frac{1}{b-a}, for a<x<ba < x < bModelling random choice where all results in an interval are equally likely.
ExponentialExp(λ)Exp(\lambda)λeλx\lambda e^{-\lambda x}, for x>0x > 0Modelling waiting times, queue times, or lifetimes.
NormalNormal(μ,σ2)Normal(\mu, \sigma^2)1σ2πe(xμ)2/2σ2\frac{1}{\sigma \sqrt{2\pi}} e^{-(x-\mu)^2 / 2\sigma^2}The fundamental distribution in statistics, arising as a limit of many phenomena.

Feature Highlight: Exponential Memoryless Property

The exponential distribution has a unique characteristic shared with the discrete Geometric distribution: the memoryless property. If XExp(λ)X \sim Exp(\lambda), the probability that the event has not occurred by time ss, and still requires at least tt more time, is the same as the probability it would have required time tt from the start: P(X>s+tX>s)=P(X>t)\mathbf{P(X > s + t | X > s) = P(X > t)}.

Standardizing and Linear Transformations

If a random variable XX has density fX(x)f_X(x) and you create a new random variable Y=aX+bY = aX + b (where a0a \neq 0), the density of YY (fY(y)f_Y(y)) is directly related to fX(x)f_X(x):

fY(y)=1afX(yba)f_Y(y) = \frac{1}{|a|} f_X \left( \frac{y-b}{a} \right)

This rule is vital for standardization, especially for the Normal distribution. If XNormal(μ,σ2)X \sim Normal(\mu, \sigma^2), the standardized variable Z=XμσZ = \frac{X-\mu}{\sigma} always results in a Standard Normal variable, ZNormal(0,1)Z \sim Normal(0, 1).

Example: Normal Probability Calculation

Q2

Cashew Weights

Example: A machine filling bags of cashews produces weights YNormal(200,42)Y \sim \text{Normal}(200, 4^2).

Question: How likely is it that a bag has fewer than 195 grams?

📝 View Detailed Solution

Solution: We standardize 195 grams to find the equivalent ZZ-score, where ZNormal(0,1)Z \sim \text{Normal}(0, 1): P(Y<195)=P(Z<1952004)=P(Z<1.25)P(Y < 195) = P \left( Z < \frac{195 - 200}{4} \right) = P(Z < -1.25)

Using the symmetry of the Normal distribution and/or a table, we find P(Z<1.25)0.106P(Z < -1.25) \approx 0.106. There is about a 10.6% chance of producing a bag this light.


5.4 Multiple Continuous Random Variables

When dealing with two or more continuous random variables (X,YX, Y), their relationship is summarized using joint, marginal, and conditional densities.

Concept: Joint and Marginal Densities

  1. Joint Density (f(x,y)f(x, y)): This function describes the likelihood of both XX and YY taking on values within a two-dimensional region AR2A \subset \mathbb{R}^2. P((X,Y)A)=Af(x,y)dxdyP((X, Y) \in A) = \iint_A f(x, y) dx dy The volume under f(x,y)f(x, y) over the entire plane must equal 1.

  2. Marginal Density (fX(x)f_X(x)): If you are only interested in the probability distribution of XX, regardless of YY‘s outcome, you find the marginal density by integrating the joint density over all possible values of YY: fX(x)=f(x,y)dyf_X(x) = \int_{-\infty}^{\infty} f(x, y) dy

Concept: Independence

Two continuous random variables XX and YY are independent if and only if their joint density is simply the product of their marginal densities for all x,yRx, y \in \mathbb{R}:

f(x,y)=fX(x)fY(y)f(x, y) = f_X(x) f_Y(y)

If XX and YY are uniformly distributed over a circular disk (a dependent region), they are generally not independent.

  • If you observe XX near the edge of the disk (e.g., X=4X=4 in a disk of radius 5), the range of possible values for YY becomes severely limited (dependent).
  • If they were independent, the joint density would factor. For a disk of radius 5, the marginal density fX(x)f_X(x) is found to be proportional to 25x2\sqrt{25 - x^2}. The joint density f(x,y)=1/(25π)f(x, y) = 1/(25\pi) does not equal the product of the marginal densities.

Concept: Conditional Density

In the continuous setting, calculating the probability of XX given that YY takes an exact value bb (P(XY=b)P(X|Y=b)) is formally impossible because P(Y=b)=0P(Y=b)=0.

Instead, we define a conditional density based on the joint and marginal densities, which represents the distribution of XX given that YY is known to be bb:

fXY=b(x)=f(x,b)fY(b)f_{X|Y=b}(x) = \frac{f(x, b)}{f_Y(b)} (Provided fY(b)>0f_Y(b) > 0).

Once the conditional density is found, you can calculate conditional probabilities by integration: P(XAY=b)=AfXY=b(x)dxP(X \in A | Y=b) = \int_A f_{X|Y=b}(x) dx

Example: Conditional Uniformity

Q3

Triangular Distribution

Example: Suppose the joint density f(x,y)f(x, y) is uniform over the region T={(x,y)0<x<y<4}T = \{(x, y) | 0 < x < y < 4\} (a triangular region). The marginal density of YY is fY(y)=y/8f_Y(y) = y/8 for 0<y<40 < y < 4.

Question: What is the conditional distribution of XX given Y=bY=b?

📝 View Detailed Solution

Solution: Using the definition: fXY=b(x)=f(x,b)fY(b)=1/8b/8=1b,for 0<x<bf_{X|Y=b}(x) = \frac{f(x, b)}{f_Y(b)} = \frac{1/8}{b/8} = \frac{1}{b}, \quad \text{for } 0 < x < b

The conditional distribution (XY=b)(X|Y=b) is Uniform(0,b)(0, b). This makes sense: if we know YY is fixed at bb, XX can be anywhere between 0 and bb.