Continuous Probabilities and Random Variables
CONTINUOUS PROBABILITIES
Chapter 5 of your statistics textbook, Continuous Probabilities and Random Variables , marks a significant transition in how we define and calculate probabilities. Unlike discrete variables (like counts or die rolls) which have a countable number of outcomes, continuous variables (like temperature, time, or height) can take on any value within a range. This shift requires abandoning sums and introducing the concept of density and integration.
5.1 The Necessary Shift: From Sums to Integrals
In discrete probability, every outcome has a specific chance of occurring, and we calculate the probability of an event by summing up the likelihoods of the relevant outcomes. This approach fails entirely in the continuous world.
Concept: Zero Probability for Individual Outcomes
If you try to assign a uniform probability to every single outcome in an uncountable interval (like (0, 1)), the total probability would diverge to infinity if .
Therefore, for any continuous random variable , the probability of landing on any single exact point is zero: .
Question: If for every point , how can the total probability still be 1?
Explanation: The fundamental axioms of probability only allow us to sum up the probabilities of a countable collection of disjoint events. Since the interval (0, 1) contains an uncountable collection of points, the probability of the entire space is still 1, even though the probability of any individual point is zero.
Concept: The Probability Density Function (PDF)
Since individual points have zero probability, we must define probability over intervals or sets, which is done using a function called the probability density function (), or simply the density.
- Non-Negative: The density must be non-negative for all values of (probabilities cannot be negative).
- Total Area is One: The total area under the density curve must equal one (representing 100% certainty that the outcome occurs somewhere on the real line): .
Probability via Integration
The probability of a continuous random variable falling into an event is found by calculating the area under the density curve over that set :
Example: Non-Uniform Density
Area under Density
A density function may not be constant, meaning equal-length intervals might have different probabilities.
If a density is given by for :
The second interval (which is further away from zero) has a much higher probability, reflecting that the density function is larger in that region.
5.2 & 5.3 Summarizing Continuous Variables
Concept: The Cumulative Distribution Function (CDF)
The distribution function or Cumulative Distribution Function () of a random variable gives the probability that is less than or equal to a specific value :
For a continuous random variable with density , the CDF is the integral of the density up to point :
This relationship is crucial, as the CDF is differentiable wherever the density is continuous, and . This means that the PDF and the CDF contain the exact same information about the distribution of .
Key Continuous Distributions
| Distribution | Notation | Density Function () | Core Application |
|---|---|---|---|
| Uniform | , for | Modelling random choice where all results in an interval are equally likely. | |
| Exponential | , for | Modelling waiting times, queue times, or lifetimes. | |
| Normal | The fundamental distribution in statistics, arising as a limit of many phenomena. |
Feature Highlight: Exponential Memoryless Property
The exponential distribution has a unique characteristic shared with the discrete Geometric distribution: the memoryless property. If , the probability that the event has not occurred by time , and still requires at least more time, is the same as the probability it would have required time from the start: .
Standardizing and Linear Transformations
If a random variable has density and you create a new random variable (where ), the density of () is directly related to :
This rule is vital for standardization, especially for the Normal distribution. If , the standardized variable always results in a Standard Normal variable, .
Example: Normal Probability Calculation
Cashew Weights
Example: A machine filling bags of cashews produces weights .
Question: How likely is it that a bag has fewer than 195 grams?
📝 View Detailed Solution ▼
Solution: We standardize 195 grams to find the equivalent -score, where :
Using the symmetry of the Normal distribution and/or a table, we find . There is about a 10.6% chance of producing a bag this light.
5.4 Multiple Continuous Random Variables
When dealing with two or more continuous random variables (), their relationship is summarized using joint, marginal, and conditional densities.
Concept: Joint and Marginal Densities
-
Joint Density (): This function describes the likelihood of both and taking on values within a two-dimensional region . The volume under over the entire plane must equal 1.
-
Marginal Density (): If you are only interested in the probability distribution of , regardless of ‘s outcome, you find the marginal density by integrating the joint density over all possible values of :
Concept: Independence
Two continuous random variables and are independent if and only if their joint density is simply the product of their marginal densities for all :
- If you observe near the edge of the disk (e.g., in a disk of radius 5), the range of possible values for becomes severely limited (dependent).
- If they were independent, the joint density would factor. For a disk of radius 5, the marginal density is found to be proportional to . The joint density does not equal the product of the marginal densities.
Concept: Conditional Density
In the continuous setting, calculating the probability of given that takes an exact value () is formally impossible because .
Instead, we define a conditional density based on the joint and marginal densities, which represents the distribution of given that is known to be :
(Provided ).
Once the conditional density is found, you can calculate conditional probabilities by integration:
Example: Conditional Uniformity
Triangular Distribution
Example: Suppose the joint density is uniform over the region (a triangular region). The marginal density of is for .
Question: What is the conditional distribution of given ?
📝 View Detailed Solution ▼
Solution: Using the definition:
The conditional distribution is Uniform. This makes sense: if we know is fixed at , can be anywhere between 0 and .