Grinstead & Snell Chapters 4-6: Probability Distributions and Entropy
Overview
Section titled “Overview”These chapters transition from discrete to continuous probability, introduce expected value and variance as key summary statistics, and culminate in Shannon entropy as a measure of uncertainty. This material is foundational for information theory.
Key Concepts Introduced
Section titled “Key Concepts Introduced”- [[Continuous Random Variables]] — Variables taking values on a continuum; described by probability density functions (PDFs) rather than probability mass functions
- [[Expected Value]] — The “center of mass” of a distribution;
- [[Variance and Standard Deviation]] — Measures of spread;
- [[Common Distributions]] — Uniform, exponential, normal, and their properties
- [[Shannon Entropy]] — The expected surprise of a random variable;
Main Results
Section titled “Main Results”Result 1: Continuous Probability Density
Section titled “Result 1: Continuous Probability Density”For a continuous random variable with PDF :
Note: for any specific value . Probability mass is “smeared” over intervals.
Result 2: Linearity of Expectation
Section titled “Result 2: Linearity of Expectation”For any random variables and (not necessarily independent):
This is surprisingly powerful—it holds even when and are dependent.
Result 3: Variance of Independent Sum
Section titled “Result 3: Variance of Independent Sum”For independent and :
Note: Independence is required here, unlike for expectation.
Important Definitions
Section titled “Important Definitions”| Term | Definition |
|---|---|
| A function with ; | |
| CDF | |
| Expected Value | (continuous) or (discrete) |
| Variance | |
| Entropy |
Key Equations
Section titled “Key Equations”| Name | Equation | Use |
|---|---|---|
| Normal PDF | Most common distribution; Central Limit Theorem | |
| Exponential PDF | for | Waiting times; memoryless property |
| Binary Entropy | Entropy of a biased coin | |
| Entropy bound | Maximum entropy is uniform |
Proofs to Remember
Section titled “Proofs to Remember”-
Entropy maximization (uniform) — Key insight: use Lagrange multipliers with normalization constraint; the uniform distribution maximizes subject to .
-
Variance decomposition — Key insight: follows from expanding and using linearity.
Connections to Dissertation
Section titled “Connections to Dissertation”- Entropy is the central concept; everything flows from here
- Expected value is used in defining entropy ( is the expected surprisal)
- Normal distribution appears in maximum entropy with fixed mean and variance
- The exponential distribution is maximum entropy with fixed mean (see [[Maxent Mean Constraint]])
Questions Raised
Section titled “Questions Raised”- [[How Does Differential Entropy Differ From Discrete Entropy?]]
- Why can differential entropy be negative while discrete entropy can’t?
Problems to Work
Section titled “Problems to Work”- Problem 4.2: Entropy of biased coin
- Problem 5.15: Expected value of geometric distribution
- Problem 6.8: Entropy of mixture distribution
- Problem 6.12: Prove entropy maximization for uniform
What I Found Difficult
Section titled “What I Found Difficult”The transition from discrete to continuous distributions requires careful handling. In particular:
- is counterintuitive; you have to think about intervals, not points
- Differential entropy can be negative, which seems to violate the intuition that “uncertainty is non-negative”
- The definition of conditional PDF requires division by a probability that equals zero… resolved via limits
What Surprised Me
Section titled “What Surprised Me”Linearity of expectation doesn’t require independence! This is incredibly useful. For example, to find the expected number of fixed points in a random permutation, you can sum indicator variables without worrying about their complex dependencies.
Next Steps
Section titled “Next Steps”- Review section on conditional expectation
- Work through the entropy exercises more carefully
- Connect to [[Cover & Thomas Chapter 2]] for information-theoretic perspective
- Build the [[Binary Entropy Explorer]] demonstration