Summarizing Discrete Random Variables
SUMMARIZING DISCRETE RVs
Chapter 4 of the “Stats 2 Book”, Summarizing Discrete Random Variables , focuses on deriving numerical characteristics—like the average and the spread—that summarize the behaviour of a random variable (RV) without listing every possible outcome. These summaries allow you to understand the “typical” result and the likelihood of unusual results.
The chapter covers five main concepts: Expected Value (Average), Variance and Standard Deviation (Spread), Conditional Expectation, and Covariance and Correlation (Relationships between RVs).
4.1 Expected Value (): The Average Outcome
The expected value (), or average, of a discrete random variable is essentially a long-run weighted average of all its possible outcomes. It tells you what value you should expect, on average, if you repeated the experiment many times.
Concept and Calculation
Expected Value Formula
The calculation of the expected value relies on weighting each possible outcome () by its likelihood () and summing these products:
For beginner intuition, consider rolling a standard fair die. The possible outcomes are , each with a probability .
Example: Expected Value of a Lottery Ticket
Lottery Valuation
Example (4.1.2): A lottery ticket can be worth £200, £20, or nothing (£0).
Question: What is the average value of such a ticket?
📝 View Detailed Solution ▼
Solution: Applying the definition of expected value:
Answer: The expected value of the ticket is £0.74 (or 74 pence).
Key Properties of Expected Value
Application: Expected Values of Common Distributions
Using these rules simplifies calculating the expectations for fundamental distributions:
| Distribution | Parameter(s) | Expected Value () |
|---|---|---|
| Bernoulli() | (Success/Failure) | |
| Binomial() | trials, success probability | |
| Geometric() | Trials until 1st success |
The expected number of successes in independent Bernoulli trials (Binomial) is simply the sum of the individual Bernoulli expected values: .
4.2 Variance and Standard Deviation: Quantifying Spread
While the expected value tells you the central point of the distribution, it does not tell you how spread out the outcomes are.
Concept and Calculation
Variance & Standard Deviation
The variance () is the average of the squared distances between each outcome and the mean .
The standard deviation ( or ) is the square root of the variance. The standard deviation is typically viewed as the typical distance from the average.
Key Properties of Variance
- Alternate Formula: Variance is often easier to compute using the second moment ():
- Scaling: When scaling a random variable by a constant , the variance scales by :
- Shifting: Adding a constant (a location shift) does not change the spread:
- Independence (Sum Rule): If and are independent random variables, their variances add:
Example: Variance of a Die Roll
Die Roll Spread
Question: What are the variance and standard deviation of a single roll of a fair die ()?
📝 View Detailed Solution ▼
Solution: The variance is the average squared distance from the mean 3.5:
The Standard Deviation is . This confirms that a typical deviation from the mean 3.5 is about 1.71.
Variance of Common Distributions
| Distribution | Variance () |
|---|---|
| Bernoulli() | |
| Binomial() | |
| Geometric() |
4.3 Standard Units and Chebyshev’s Inequality
When discussing variability, it is useful to express outcomes in standard units—the number of standard deviations () an outcome is from the expected value ().
Chebyshev’s Inequality
The Chebyshev’s Inequality provides a universal upper bound on the probability that a random variable will deviate significantly from its mean, regardless of the specific shape of the distribution.
Chebyshev's Inequality
For any random variable with finite variance , and for any value :
Explanation: This inequality states that the chance of an outcome being or more standard deviations away from the average is at most .
Example (4.3.5)
Chebyshev Bound
Example (4.3.5): Find an upper bound on the likelihood that will be more than two standard deviations from its expected value.
📝 View Detailed Solution ▼
Solution: Set :
Conclusion: There is at most a 25% chance that a random variable will be more than two standard deviations from its expected value.
4.4 Conditional Expectation and Total Expectation
Conditional expected value () is the expected value of given that some event has definitely occurred. The conditioning event alters the underlying probabilities, thus potentially changing the average outcome.
Law of Total Expectation
This theorem relates the overall expected value to conditional expected values over a set of disjoint events that cover the entire sample space ().
Law of Total Expectation
Example (4.4.5)
Investment Return
Example (4.4.5): A venture capitalist estimates the expected return on an investment () conditional on economic outlooks (stronger), (same), or (weaker):
- million (P(A) = 0.1)
- million (P(B) = 0.4)
- million (P(C) = 0.5)
Question: What is the overall expected return ()?
📝 View Detailed Solution ▼
Solution: Using the Law of Total Expectation:
The expected return on investment is $0.2 million (£200,000).
4.5 Covariance and Correlation: Measuring Relationships
When analyzing two random variables, and , the goal is often to determine how they relate to each other—if knowing one changes your expectation of the other.
Covariance ()
The covariance measures the degree to which two random variables vary together.
- Positive Covariance (): tends to be above average when is above average. They are positively correlated.
- Negative Covariance (): tends to be above average when is below average. They are negatively correlated.
- Zero Covariance (): and are uncorrelated.
Property: If and are independent, then . (The converse is not necessarily true).
The calculation is often simplified using the alternate formula:
Correlation ()
The correlation coefficient () is the dimensionless, standardized version of the covariance. It is standardized by dividing the covariance by the product of the standard deviations ().
Correlation Coefficient
Range and Interpretation: The correlation coefficient is always bounded between and .
- : Strong positive linear relationship.
- : Strong negative linear relationship.